Abstract
Graph Neural Networks (GNNs) have emerged as powerful tools for predicting material properties, yet they often struggle to capture many-body interactions and require extensive manual feature engineering. Here, we present EOSnet (Embedded Overlap Structures for Graph Neural Networks), a novel approach that addresses these limitations by incorporating Gaussian Overlap Matrix (GOM) fingerprints as node features within the GNN architecture. Unlike models that rely on explicit angular terms or human-engineered features, EOSnet efficiently encodes many-body interactions through orbital overlap matrices, providing a rotationally invariant and transferable representation of atomic environments. The model demonstrates superior performance across various prediction tasks of materials’ properties, achieving particularly notable results in properties sensitive to many-body interactions. For band gap prediction, EOSnet achieves a mean absolute error of 0.163 eV, surpassing previous state-of-the-art models. The model also excels in predicting mechanical properties and classifying materials, with 97.7% accuracy in metal/nonmetal classification. These results demonstrate that embedding GOM fingerprints into node features enhances the ability of GNNs to capture complex atomic interactions, making EOSnet a powerful tool for materials’ discovery and property prediction.


Machine learning (ML) has become an indispensable tool in the fields of materials science and condensed matter physics, enabling the efficient prediction and discovery of material properties that would otherwise be time-consuming using first-principles calculations. As the demand for high-throughput materials screening has grown, ML models, particularly deep learning architectures, , have evolved to handle increasingly complex data sets, capturing intricate relationships between atomic structures and their properties. Within this context, Graph Neural Networks (GNNs) , have emerged as particularly well-suited for modeling atomic and molecular systems due to their ability to process graph-based data, where atoms and their interactions are represented as nodes and edges, respectively.
In the early applications of ML to materials science, models such as high-dimensional neural networks (HDNNs) were employed to learn the complex potential energy surface (PES) across large atomic systems. These models utilized various descriptors like Atom-Centered Symmetry Functions (ACSFs), , SO(3) power spectrum or Smooth Overlap of Atomic Positions (SOAP), , SO(4) bispectrum or Spectral Neighbor Analysis Potential (SNAP), and Atomic Cluster Expansion (ACE). − Each descriptor had its own trade-offs. For example, ACSFs required parameter tuning, while SOAP and SNAP were computationally intensive due to spherical harmonics and higher-order angular terms. , ACE provided a complete, scalable framework for atomic representation but introduced challenges with higher-order term computations.
HDNNs demonstrated near-DFT accuracy in predicting energy-related properties for elemental crystals with simple lattice types. However, these models require individual training for specific chemical systems and struggle with binary, ternary, or quaternary systems featuring complex lattices, limiting their transferability. Additionally, all four descriptors focused on local atomic environments within a cutoff radius, lacking a global encoding of the entire systemic geometry and topology. Efforts to incorporate long-range interactions, such as electrostatics , and nonlocal charge-transfer, − has greatly improved HDNN accuracy but still primarily targeted DFT energies and partial charges. Despite these advancements, accurate prediction of electronic properties, such as Fermi energy and band gap, remains an area requiring further exploration.
GNNs, on the other hand, eliminate the need for extensive descriptor engineering by directly learning from structural data. This capability enables GNNs to model both the local atomic environment and the global structure of materials in a more natural and scalable manner, making them particularly useful for applications in materials discovery, such as predicting electronic properties, phase transitions, and mechanical properties. − One of the most influential models in this area is the Crystal Graph Convolutional Neural Network (CGCNN), introduced by Xie and Grossman in 2018. CGCNN represents materials as graphs to capture the structural and bonding information on crystalline solids. This representation, combined with convolutional layers that update atomic and bond features through a message passing mechanism, makes CGCNN highly effective in predicting material properties from crystal structures. The success of CGCNN sparked widespread interest in the use of GNNs for material property prediction.
Despite the success, distance-based message-passing GNN architectures like CGCNN face two major limitations. One limitation is that they only encode pairwise distance information for edge features, neglecting many-body interactions that are essential for accurately modeling atomic environments. Many-body interactions involve complex dependencies among multiple atoms, where the properties of an atom cannot be fully described by its interactions with just one or two neighbors. Another challenge is that current GNN-based machine learning models often require extensive human intervention and numerous in silico experiments to select appropriate chemical information (such as atomic number, valence electrons, electronegativity, covalent radius) for node feature encoding. This manual feature selection limits the model performance and transferability, particularly when working with smaller data sets. ,
To address the first limitation, researchers have enhanced GNN architectures by incorporating extra geometric information and improving how these models process atomic interactions. One key improvement involves integrating more complex geometric features beyond simple bond lengths, such as angular and directional information. Models like DimeNet, GemNet, ALIGNN, and M3GNet leverage these geometric details to capture intricate spatial relationships between atoms, which is particularly important in anisotropic materials. Moreover, models like E3NN and NequIP incorporate vector information to ensure the representations remain invariant under rotations and translations, thereby enhancing the ability to model complex atomic systems. In addition, attention mechanisms and global aggregation techniques have revolutionized how GNNs handle information flow. Attention layers, used in models like GATGNN and Equiformer, dynamically adjust the importance of different atomic interactions, allowing the network to emphasize the most chemically relevant bonds. These attention mechanisms enhance local expressiveness and improve the aggregation of global structural information. By combining these features with advanced readout functionssuch as global attention and hierarchical poolingmodels like MEGNet and GraphTrans ensure that critical information is preserved across the entire structure. Together, these innovations allow GNNs to efficiently prioritize and synthesize both local and global atomic interactions, improving the accuracy of material property predictions. Beyond these general developments, there have also been efforts to tailor GNNs specifically for crystalline materials. For example, GeoCGNN and Matformer explicitly encode crystal periodicity, while the Reciprocal Space Neural Network (RSNN) leverages reciprocal space information to capture long-range interactions in periodic systems. These developments underscore the growing complexity and sophistication of GNN architectures in materials science, enabling more accurate and generalizable property predictions. Recently, MACE , has been proposed as another advanced framework that systematically incorporates many-body interactions into message-passing neural networks by employing high-order tensor products of spherical harmonics. MACE has demonstrated state-of-the-art performance for atomic-scale simulations, primarily excelling in fitting potential energy surfaces for molecules and materials.
In this work, we introduce EOSnet (Embedded Overlap Structures for Graph Neural Networks) to overcome the limitations in GNN models for materials science. EOSnet adopts a new approach to incorporate many-body interactions through Gaussian Overlap Matrix (GOM) fingerprints , as atomic features within the GNN architecture. Unlike previous GNN models that embed angular information into edge features explicitly, or rely on spherical harmonics expansions and careful truncation to achieve rotational invariance and many-body completeness (as in MACE, , which is mainly for fitting potential energy surfaces), EOSnet focuses on predicting material properties, such as electronic band gaps. By representing the overlap of atomic orbitals between neighboring atoms through GOMs, EOSnet provides a compact and efficient description of the many-body interactions that govern these properties. Moreover, the fingerprint features in EOSnet are generic and transferable, ensuring that two atoms share the same node feature only if they have identical neighboring environments, even within the same atomic species. Another advantage of the GOM-based fingerprints is their inherent rotational and translational invariance, ensuring a consistent description of atomic environments regardless of structural orientation or position. By embedding these fingerprints as node features, EOSnet achieves superior predictive performance across a range of material property tasks, particularly those where orbital overlaps are critical, while maintaining computational efficiency.
In EOSnet, the crystal structure is represented as a graph, where each atom is a node, and the edges represent bonds between neighboring atoms. The model architecture, depicted in Figure , integrates GOM-based fingerprints within a graph convolutional neural network (GCNN) framework. These GOM fingerprints are derived from the eigenvalues of Gaussian overlap matrices, designed to capture many-body atomic interactions within a cutoff sphere centered around each atom in the crystal structure.
1.

Schematic architecture of EOSnet, demonstrating how GOM Fingerprints are embedded into the Graph Convolutional Neural Network (GCNN) framework. Here we use the SrTiO3 crystal structure (top-down view) as an example. Each atom in the crystal are used to build the atom-centered Gaussian Overlap Matrix (GOM) that captures neighbor orbital overlaps. A normalized GOM for an O atom is shown as an example. The diagonal elements represent self-overlap of atoms, with the (1,1) element corresponding to the central atom self-overlap, which is maximal due to the smooth cutoff function f c used in GOM construction. The first row/column shows overlaps between the central atom and its neighbors, while other off-diagonal elements show overlaps between neighbors. The GOM eigenvalues are input as original node features (purple) for the GCNN. Then the CGNN learn the hidden node features by feature aggregation and message-passing (see details in Figure ) in the graph convolution (orange). Then hidden layers (cyan) are used to extract hidden features of each node then use average pooling to pool the node features to crystal feature (pink). Finally we use on last fully connected layer (FCN) to produce the predicted material property, denoted by α.
To construct the GOM, Gaussian-type orbitals (GTOs) are centered on each atom within this sphere, and the overlap integrals are calculated between every pair of atoms within the cutoff radius. The width of each Gaussian function is determined by the covalent radius of the atom on which it is centered, incorporating key features of the covalent radius into the atomic representation. Additionally, a cutoff function ensures that the overlap integrals smoothly decay to zero at the boundary of the sphere, preventing discontinuities when atoms enter or leave the region. The definitions of the GOM matrix elements and the smooth cutoff function are given by
| 1a |
| 1b |
where O m represents the GOM for atom m in the unit cell, and f c is the smooth cutoff function, defined by a user-specified cutoff radius r cut and a nonlinearity parameter n. Indices i and j label the atoms inside the cutoff sphere, and r im =|r i -r m | and r jm =|r j -r m | denote the distances between atom m and atoms i and j, respectively. The orbital indices l and l′ correspond to specific angular momentum states. The Gaussian-type orbitals (GTOs) are given by
where N l is the normalization factor and , with r i being the covalent radius of atom i.
The fingerprint for each atom is then computed by extracting the eigenvalues of its corresponding GOM. These eigenvalues encode essential information about atomic environments and serve as rotation and translation invariant features for the graph neural network. This design captures the strength of interactions between an atom and its neighbors, as well as the interactions among the neighboring atoms themselves, offering a comprehensive and efficient representation of many-body atomic interactions. Since the Gaussian parameters used in the GTOs are fixed, the descriptor construction process is simplified, avoiding extra tuning for specific systems. For moderate-sized systems, GOM-based fingerprints offers a balance between expressiveness and computational efficiency. Unlike methods that rely on costly angular expansions, such as SOAP and bispectrum descriptors, GOMs produce a matrix (or eigenvalue) representation that can be directly fed into machine learning models without extensive preprocessing. This straightforward mathematical structure allows neural networks to capture both local and global structural properties effectively. Figure illustrates the message-passing mechanism of EOSnet and the many-body interaction nature of the GOM fingerprints in a cubic SrTiO3 structure, illustrating how the overlap strength between atomic orbitals is encoded in the GOM.
2.

Illustration of many-body interaction and GOM. Left: Demonstration of the feature aggregation and message-passing scheme for an SrTiO3 crystal using GOM fingerprints. Green, blue, and red spheres represent Sr, Ti, and O atoms, respectively. For simplicity, only four Sr atoms are considered as nearest neighbors to the Ti atom. The m (i,j) is the message from atom j to atom i denoted in the Equation in the main article. n i denotes the node feature of atom i at layer l, e (i,j) represent the edge feature between atom i and j from Equation to (). Right: Normalized Gaussian Overlap Matrix for an O atom, with atoms within the cutoff radius sorted by their distance from the central atom. The diagonal elements indicates the self-overlapping, and the first entry of the diagonal elements correspond to the central atom. For demonstration purposes, we only show the first 16 GOM overlap elements with s orbitals.
The EOSnet architecture follows a standard message-passing framework, where the node features (GOM fingerprints) are updated based on the interactions between neighboring atoms. The workflow of EOSnet is illustrated in Figure . In each convolutional layer, the node feature for atom i is updated by aggregating information from its neighboring atoms j through a message-passing mechanism. This process is defined as
| 1 |
| 2 |
3.
Workflow of EOSnet, which uses fractional coordinates ( , where ), lattice vectors , atomic numbers (Z), and covalent radii (R cov) as inputs. The denotes bond distance expansion using Gaussian filters, defined in eq . The process incorporates atomic embedding, many-body interaction, and bond distance expansion to generate node and edge features, which are processed through graph convolution layers and pooling to predict material properties.
Here, is the message from atom j to atom i via the k-th edge, n i denotes the node feature of atom i at layer l, e (i,j) k represent the k-th edge feature between atom i and j. And W g , b g are the gate weight and bias at layer l, W m , b m are the message weight and bias at layer l. Operator ⊙ is the element-wise matrix multiplication, ⊕ is the vector concatenation, σ and g are independent nonlinear activation functions. EOSnet adopts a gated architecture similar to that introduced in the CGCNN model. A simple convolution with a shared weight matrix W treats all neighbors identically, neglecting differences in their interaction strengths. By introducing a gate function σ and another nonlinear transformation g, it creates a multiplicative interaction that allows the network to selectively control how much information from each neighbor is incorporated.
The convolutional layers, as depicted in Figure , process the GOM-embedded node features, allowing the network to retain information about the many-body interactions throughout the depth of the model. After each message-passing step, node features are updated using gated convolutional layers. These layers ensure that relevant information from the neighbors is retained while irrelevant information is filtered out. This process is repeated for several layers, allowing the model to capture both local and global atomic interactions.
At the end of the graph convolutional layers, the node features are aggregated using a global pooling operation to predict the target property. In EOSnet, we use an average pooling function, which computes the mean of all node features in the graph, followed by a fully connected layer that maps the aggregated feature vector to the target property:
| 3 |
where N at is the number of atoms in the crystal structure, and is the final node feature after the last convolutional layer. The fully connected layer (FCN) maps the aggregated node features to the target property, providing the final prediction. This averaging operation occurs between the hidden layers and the output layer in GCNN, as shown in Figure . After the graph convolution layers have generated hidden features for each atom, an average pooling step aggregates these atom-level embeddings into a single vector that represents the entire crystal. This pooled crystal-level feature is then passed to the fully connected layers to produce the final property prediction. The a i in the input layer of Figure represents the fingerprint-embedded atomic descriptor for the i-th atom in the crystal. Each a i is derived from the GOM eigenvalues.
In addition to the GOM fingerprints used for node features, EOSnet also utilizes bond distance expansions to define the edge features between atoms. These edge features are expanded using Gaussian filters to ensure smooth and continuous representation of interatomic distances. The Gaussian distance expansion of the k-th bond is expressed as
| 4 |
where ∥r i,j ∥ is the distance between atoms i and j, d k is the center of the Gaussian filter and σ k the scale parameter controlling the Gaussian width. These expanded bond distances provide additional geometric information for the convolutional layers, enabling the model to effectively learn from both node and edge features. The inclusion of these features is crucial for capturing short-range interactions, especially in systems where the geometric arrangement of atoms significantly impacts material properties. The workflow for this process is also detailed in Figure .
EOSnet is trained on a variety of material data sets, including both large (>20,000) and small (<2,000) data sets for material property prediction. The model is trained using a supervised learning approach with an 80%–10%–10% train-validation-test split. Mean absolute error (MAE) is used as the loss function for regression tasks, such as predicting formation energy, total energy, and bandgap. For classification tasks, such as distinguishing between metals and nonmetals, cross-entropy loss is used. A dynamic class weight scheme is implemented to eliminate the effect of class imbalance in binary classification problems. The model is optimized using the Adam optimizer with a learning rate scheduler to dynamically adjust the learning rate during training. Gradient clip technique is used to prevent overfitting during training process.
The performance of the proposed EOSnet model was evaluated across various material property prediction tasks, including the prediction of total energy, formation energy, Fermi energy, band gap, bulk modulus, shear modulus, and a metal/nonmetal classification task. To benchmark the effectiveness of EOSnet, comparisons were made with several models, including CGCNN, MEGNet, and M3GNet. The summarized results are presented in Table . It is important to note that although all models listed in Table were trained on data sets from the Materials Project, the sizes of the data sets and the evaluation metrics may differ. The primary focus of this comparison is not to claim that EOSnet is universally superior but to highlight its performance across various tasks. Specifically, we emphasize the advancements made by EOSnet in capturing many-body interactions to enhance the prediction accuracy of material properties. Studies show that the MAE of ML models generally decreases with the increase in the size of the training data set. ,, Therefore, the performance of EOSnet is expected to improve further with larger training data sets in the future studies.
1. Comparison of the MAEs in Total Energy (E tot), Formation Energy (E form), Fermi Energy (E f ), Band Gap (E g ), Bulk Modulus (K VRH), Shear Modulus (G VRH) and Metal/Nonmetal Classification between EOSnet and Previous Works .
| EOSnet | CGCNN | M3GNet | MEGNet | |
|---|---|---|---|---|
| Etot (eV ·atom–1) | 0.034 | 0.072 | 0.035 | - |
| (19,364) | (28,046) | (132,752) | ||
| Eform (eV ·atom–1) | 0.022 | 0.039 | 0.0195 | 0.028 |
| (131,240) | (28,046) | (132,752) | (60,000) | |
| E f (eV) | 0.295 | 0.363 | - | - |
| (27,293) | (28,046) | |||
| E g (eV) | 0.163 | 0.388 | 0.183 | 0.33 |
| (19,393) | (16,485) | (106,113) | (36,720) | |
| KVRH (log10 (GPa)) | 0.034 | 0.054 | 0.058 | 0.050 |
| (5,000) | (2,041) | (10,987) | (4,664) | |
| GVRH (log10 (GPa)) | 0.072 | 0.087 | 0.086 | 0.079 |
| (5,000) | (2,041) | (10,987) | (4,664) | |
| Metal/Nonmetal Classifier | 97.7% | 95.0% | 95.8% | 90.6% |
| (19,393) | (16,458) | (106,113) | (55,391) |
The number of structures in the training data is in parentheses.
For our model, the numbers in parentheses indicate the total numbers of structures before train-validation-test split. 19,364 E tot data are from ASE cubic perovskite data set. − 27,293 E f data and 19,393 band gap data are from Materials Project 2024 data set. 5,000 Voigt-Reuss-Hill (VRH) average bulk moduli, shear moduli are from Materials Project 2018 data set, and 131,240 E form data are from Materials Project 2019 data set.
Figure a and b provide a detailed analysis of the EOSnet model performance in predicting formation energy and electronic band gap. The color gradient in these figures represents the density of data points, with brighter colors indicating regions of higher density. The tight clustering of predicted values around the diagonal line indicates that EOSnet effectively models the distribution of formation energy and band gap values, demonstrating robustness and reliability. An example demonstrating the strengths of EOSnet is its band gap prediction, achieving a low MAE of 0.163 eV. This performance is impressive comparing to other models, despite differences in the size of training data sets used. Band gap prediction is particularly challenging due to its sensitivity to many-body interactions, as well as electronic correlations within the material. Models that primarily rely on bond distances and basic geometric features often struggle to capture the nuanced effects that influence the electronic structure. The use of GOM-based fingerprints in EOSnet enables a more comprehensive representation of many-body interactions. This method captures not only the local interactions between an atom and its immediate neighbors but also the interactions among neighboring atoms themselves. For example, in semiconducting materials, the band gap, defined by the energy separation between valence and conduction bands, emerges from complex interplays between atomic arrangements, chemical bonding, and the electronic environment. While conventional GNNs often struggle to capture subtle electronic interactions and intricate orbital overlaps that determine band gaps, EOSnet effectively encodes these crucial quantum mechanical features.
4.
Evaluation for regression and binary classification tasks on the performance of our EOSnet model. (a) Parity plot for formation energy predictions on 131,240 data points from the Materials Project, with MAE of 0.022 eV/atom. (b) Parity plot for the prediction of electronic band gap (E g ) using 19,393 data points from the Materials Project, showing an MAE of 0.163 eV. (c) ROC curve for metal/nonmetal classification, achieving an AUC of 0.977. (d) ROC curve for dynamically stable/unstable classification on 1,335 guest-atom-substituted type-VII boron-carbide clathrates (MB6–x C x , x from 1 to 5), achieving an AUC of 0.945. The clathrate stability is determined using phonon frequency from VASP , calculations.
The enhanced performance extends beyond band gap prediction to other properties sensitive to many-body interactions, such as bulk and shear moduli (Table ). This broad success demonstrates the effectiveness of EOSnet approach to capturing complex atomic interactions in materials. The model also excels in classification tasks, achieving 97.7% accuracy in metal/nonmetal classification and 95.0% accuracy in predicting the dynamic stability of guest-atom substituted type-VII boron-carbide clathrates (Figure c and d). Particularly noteworthy is the model strong performance in stability classification despite limited training data, highlighting how the nonhuman-engineered atomic features can extract meaningful patterns even from small data sets.
In summary, we have introduced EOSnet (Embedded Overlap Structures for Graph Neural Networks), a new approach that enhances the predictive capabilities of graph neural networks in materials science by efficiently incorporating many-body interactions through GOM-based fingerprints. EOSnet provides a rotationally invariant and computationally efficient method to represent the full spectrum of atomic interactions without the need for explicit higher-order terms. Our extensive evaluations across various material property prediction tasks demonstrate the performance of EOSnet, particularly in predicting properties that are sensitive to many-body interactions such as electronic band gap and elasticity. Notably, EOSnet achieved a MAE of 0.163 eV in band gap prediction, surpassing the original CGCNN model and M3GNet. Same performance boost has been observed in the case of bulk modulus and shear modulus. This indicate embedding GOM fingerprints into node features can indeed helps the GNN model understand these properties better with the essential information on atomic environments, including the collective behavior of neighboring atoms and the strength of their orbital overlaps. This improved performance is especially important for applications in material discovery, making it a valuable tool for the design of materials with optimized electronic, mechanical, and thermal properties. While EOSnet shows performance improvements, further work could include expanding the model to handle more diverse data sets and incorporating additional geometric or attention mechanisms to capture long-range interactions more effectively. Additionally, exploring its applicability to other domains, such as catalysis or battery materials, could open new avenues for material discovery.
Supplementary Material
Acknowledgments
This work was supported by the National Science Foundation, Division of Materials Research (NSF-DMR) under Grant No. 2226700, and startup funds of the office of the Dean of SASN of Rutgers University-Newark. The authors acknowledge the Office of Advanced Research Computing (OARC) at Rutgers, The State University of New Jersey, for providing access to the Amarel cluster and associated research computing resources that have contributed to the results reported here. We thank Ying Sun for providing the training set of guest-atom substituted type-VII boron-carbide clathrates.
The source code and data for the EOSnet model are available at the following GitHub repository: https://github.com/Rutgers-ZRG/EosNet.
The Supporting Information is available free of charge at https://pubs.acs.org/doi/10.1021/acs.jpclett.4c03179.
Additional details on the model architecture, and parameters for the model configuration. It includes additional figures that further illustrate the model performance across various material property prediction tasks. eq shows Gaussian-type orbital overlap elements for Gaussian Overlap Matrix (GOM); Figures S1–S3 showing the comparison of performance of the EOSnet model with and without GOM Fingerprints for predicting various properties; Figure S4 showing the evaluation of the EOSnet model performance on formation energy predictions across varying data set sizes; Figure S5 is the histogram showing the distribution of elements per unit cell in different data set sizes; Table S1 summarizing EOSnet model parameters and training settings (PDF)
The authors declare no competing financial interest.
References
- Artrith N., Urban A.. An Implementation of Artificial Neural-Network Potentials for Atomistic Materials Simulations: Performance for TiO2. Comput. Mater. Sci. 2016;114:135–150. doi: 10.1016/j.commatsci.2015.11.047. [DOI] [Google Scholar]
- Unke O. T., Meuwly M.. PhysNet: A Neural Network for Predicting Energies, Forces, Dipole Moments and Partial Charges. J. Chem. Theory Comput. 2019;15:3678–3693. doi: 10.1021/acs.jctc.9b00181. [DOI] [PubMed] [Google Scholar]
- Wu Z., Pan S., Chen F., Long G., Zhang C., Yu P. S.. A Comprehensive Survey on Graph Neural Networks. IEEE Transactions on Neural Networks and Learning Systems. 2021;32:4–24. doi: 10.1109/TNNLS.2020.2978386. [DOI] [PubMed] [Google Scholar]
- Reiser P., Neubert M., Eberhard A., Torresi L., Zhou C., Shao C., Metni H., van Hoesel C., Schopmans H., Sommer T., Friederich P.. Graph Neural Networks for Materials Science and Chemistry. Commun. Mater. 2022;3:1–18. doi: 10.1038/s43246-022-00315-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Behler J.. Four Generations of High-Dimensional Neural Network Potentials. Chem. Rev. 2021;121:10037–10072. doi: 10.1021/acs.chemrev.0c00868. [DOI] [PubMed] [Google Scholar]
- Behler J., Parrinello M.. Generalized Neural-Network Representation of High-Dimensional Potential-Energy Surfaces. Phys. Rev. Lett. 2007;98:146401. doi: 10.1103/PhysRevLett.98.146401. [DOI] [PubMed] [Google Scholar]
- Gastegger M., Schwiedrzik L., Bittermann M., Berzsenyi F., Marquetand P.. wACSFWeighted Atom-Centered Symmetry Functions as Descriptors in Machine Learning Potentials. J. Chem. Phys. 2018;148:241709. doi: 10.1063/1.5019667. [DOI] [PubMed] [Google Scholar]
- Bartók A. P., Payne M. C., Kondor R., Csányi G.. Gaussian Approximation Potentials: The Accuracy of Quantum Mechanics, without the Electrons. Phys. Rev. Lett. 2010;104:136403. doi: 10.1103/PhysRevLett.104.136403. [DOI] [PubMed] [Google Scholar]
- Bartók A. P., Kondor R., Csányi G.. On Representing Chemical Environments. Phys. Rev. B. 2013;87:184115. doi: 10.1103/PhysRevB.87.184115. [DOI] [Google Scholar]
- Thompson A. P., Swiler L. P., Trott C. R., Foiles S. M., Tucker G. J.. Spectral Neighbor Analysis Method for Automated Generation of Quantum-Accurate Interatomic Potentials. J. Comput. Phys. 2015;285:316–330. doi: 10.1016/j.jcp.2014.12.018. [DOI] [Google Scholar]
- Drautz R.. Atomic Cluster Expansion for Accurate and Transferable Interatomic Potentials. Phys. Rev. B. 2019;99:014104. doi: 10.1103/PhysRevB.99.014104. [DOI] [Google Scholar]
- Lysogorskiy Y., van der Oord C., Bochkarev A., Menon S., Rinaldi M., Hammerschmidt T., Mrovec M., Thompson A., Csányi G., Ortner C., Drautz R.. Performant Implementation of the Atomic Cluster Expansion (PACE) and Application to Copper and Silicon. npj Comput. Mater. 2021;7:1–12. doi: 10.1038/s41524-021-00559-9. [DOI] [Google Scholar]
- Bochkarev A., Lysogorskiy Y., Menon S., Qamar M., Mrovec M., Drautz R.. Efficient Parametrization of the Atomic Cluster Expansion. Phys. Rev. Mater. 2022;6:013804. doi: 10.1103/PhysRevMaterials.6.013804. [DOI] [Google Scholar]
- Lysogorskiy Y., Bochkarev A., Mrovec M., Drautz R.. Active Learning Strategies for Atomic Cluster Expansion Models. Phys. Rev. Mater. 2023;7:043801. doi: 10.1103/PhysRevMaterials.7.043801. [DOI] [Google Scholar]
- Imbalzano G., Anelli A., Giofré D., Klees S., Behler J., Ceriotti M.. Automatic Selection of Atomic Fingerprints and Reference Configurations for Machine-Learning Potentials. J. Chem. Phys. 2018;148:241730. doi: 10.1063/1.5024611. [DOI] [PubMed] [Google Scholar]
- Caro M. A.. Optimizing Many-Body Atomic Descriptors for Enhanced Computational Performance of Machine Learning Based Interatomic Potentials. Phys. Rev. B. 2019;100:024112. doi: 10.1103/PhysRevB.100.024112. [DOI] [Google Scholar]
- Zuo Y., Chen C., Li X., Deng Z., Chen Y., Behler J., Csányi G., Shapeev A. V., Thompson A. P., Wood M. A., Ong S. P.. Performance and Cost Assessment of Machine Learning Interatomic Potentials. J. Phys. Chem. A. 2020;124:731–745. doi: 10.1021/acs.jpca.9b08723. [DOI] [PubMed] [Google Scholar]
- Dusson G., Bachmayr M., Csányi G., Drautz R., Etter S., van der Oord C., Ortner C.. Atomic Cluster Expansion: Completeness, Efficiency and Stability. J. Comput. Phys. 2022;454:110946. doi: 10.1016/j.jcp.2022.110946. [DOI] [Google Scholar]
- Yao K., Herr J. E., Toth D. W., Mckintyre R., Parkhill J.. The TensorMol-0.1 Model Chemistry: A Neural Network Augmented with Long-Range Physics. Chem. Sci. 2018;9:2261–2269. doi: 10.1039/C7SC04934J. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Faraji S., Ghasemi S. A., Rostami S., Rasoulkhani R., Schaefer B., Goedecker S., Amsler M.. High Accuracy and Transferability of a Neural Network Potential through Charge Equilibration for Calcium Fluoride. Phys. Rev. B. 2017;95:104105. doi: 10.1103/PhysRevB.95.104105. [DOI] [Google Scholar]
- Faraji S., Ghasemi S. A., Parsaeifard B., Goedecker S.. Surface Reconstructions and Premelting of the (100) CaF2 Surface. Phys. Chem. Chem. Phys. 2019;21:16270–16281. doi: 10.1039/C9CP02213A. [DOI] [PubMed] [Google Scholar]
- Becke A. D.. A Multicenter Numerical Integration Scheme for Polyatomic Molecules. J. Chem. Phys. 1988;88:2547–2553. doi: 10.1063/1.454033. [DOI] [Google Scholar]
- Xie X., Persson K. A., Small D. W.. Incorporating Electronic Information into Machine Learning Potential Energy Surfaces via Approaching the Ground-State Electronic Energy as a Function of Atom-Based Electronic Populations. J. Chem. Theory Comput. 2020;16:4256–4270. doi: 10.1021/acs.jctc.0c00217. [DOI] [PubMed] [Google Scholar]
- Schütt K. T., Sauceda H. E., Kindermans P.-J., Tkatchenko A., Müller K.-R.. SchNet – A Deep Learning Architecture for Molecules and Materials. J. Chem. Phys. 2018;148:241722. doi: 10.1063/1.5019779. [DOI] [PubMed] [Google Scholar]
- Chen C., Ye W., Zuo Y., Zheng C., Ong S. P.. Graph Networks as a Universal Machine Learning Framework for Molecules and Crystals. Chem. Mater. 2019;31:3564–3572. doi: 10.1021/acs.chemmater.9b01294. [DOI] [Google Scholar]
- Choudhary K., DeCost B.. Atomistic Line Graph Neural Network for Improved Materials Property Predictions. npj Comput. Mater. 2021;7:1–8. doi: 10.1038/s41524-021-00650-1. [DOI] [Google Scholar]
- Geiger, M. ; Smidt, T. . e3nn: Euclidean Neural Networks. arXiv, July 18, 2022, ver. 1. 10.48550/arXiv.2207.09453 [DOI]
- Yan, K. ; Liu, Y. ; Lin, Y. ; Ji, S. . Periodic Graph Transformers for Crystal Material Property Prediction. 36th Conference on Neural Information Processing Systems. 2022; pp 15066–15080, Published by Curran Associates, Inc. [Google Scholar]
- Schütt K. T., Hessmann S. S. P., Gebauer N. W. A., Lederer J., Gastegger M.. SchNetPack 2.0: A Neural Network Toolbox for Atomistic Machine Learning. J. Chem. Phys. 2023;158:144801. doi: 10.1063/5.0138367. [DOI] [PubMed] [Google Scholar]
- Gurunathan R., Choudhary K., Tavazza F.. Rapid Prediction of Phonon Structure and Properties Using the Atomistic Line Graph Neural Network (ALIGNN) Phys. Rev. Mater. 2023;7:023803. doi: 10.1103/PhysRevMaterials.7.023803. [DOI] [Google Scholar]
- Xie T., Grossman J. C.. Crystal Graph Convolutional Neural Networks for an Accurate and Interpretable Prediction of Material Properties. Phys. Rev. Lett. 2018;120:145301. doi: 10.1103/PhysRevLett.120.145301. [DOI] [PubMed] [Google Scholar]
- Gilmer, J. ; Schoenholz, S. S. ; Riley, P. F. ; Vinyals, O. ; Dahl, G. E. . Neural Message Passing for Quantum Chemistry. Proceedings of the 34th International Conference on Machine Learning. 2017; pp 1263–1272, Published by PMLR. [Google Scholar]
- Parsaeifard B., Goedecker S.. Manifolds of Quasi-Constant SOAP and ACSF Fingerprints and the Resulting Failure to Machine Learn Four-Body Interactions. J. Chem. Phys. 2022;156:034302. doi: 10.1063/5.0070488. [DOI] [PubMed] [Google Scholar]
- Acharya, D. B. ; Zhang, H. . Feature Selection and Extraction for Graph Neural Networks. Proceedings of the 2020 ACM Southeast Conference. Published by Association for Computing Machinery (ACM), New York, NY, USA, 2020; pp 252–255. [Google Scholar]
- Dwivedi V. P., Joshi C. K., Luu A. T., Laurent T., Bengio Y., Bresson X.. Benchmarking Graph Neural Networks. Journal of Machine Learning Research. 2023;24:1–48. [Google Scholar]
- Gasteiger, J. ; Groß, J. ; Günnemann, S. . Directional Message Passing for Molecular Graphs. International Conference on Learning Representations (ICLR). 2020; Published by OpenReview.net. [Google Scholar]
- Gasteiger, J. ; Becker, F. ; Günnemann, S. . GemNet: Universal Directional Graph Neural Networks for Molecules. Conference on Neural Information Processing Systems (NeurIPS). 2021; pp 6790–6802, Published by Curran Associates, Inc.. [Google Scholar]
- Choudhary K., DeCost B.. Atomistic Line Graph Neural Network for Improved Materials Property Predictions. npj Comput. Mater. 2021;7:1–8. doi: 10.1038/s41524-021-00650-1. [DOI] [Google Scholar]
- Chen C., Ong S. P.. A Universal Graph Deep Learning Interatomic Potential for the Periodic Table. Nat. Comput. Sci. 2022;2:718–728. doi: 10.1038/s43588-022-00349-3. [DOI] [PubMed] [Google Scholar]
- Batzner S., Musaelian A., Sun L., Geiger M., Mailoa J. P., Kornbluth M., Molinari N., Smidt T. E., Kozinsky B.. E(3)-Equivariant Graph Neural Networks for Data-Efficient and Accurate Interatomic Potentials. Nat. Commun. 2022;13:2453. doi: 10.1038/s41467-022-29939-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Louis S.-Y., Zhao Y., Nasiri A., Wang X., Song Y., Liu F., Hu J.. Graph Convolutional Neural Networks with Global Attention for Improved Materials Property Prediction. Phys. Chem. Chem. Phys. 2020;22:18141–18148. doi: 10.1039/D0CP01474E. [DOI] [PubMed] [Google Scholar]
- Liao, Y.-L. ; Smidt, T. . Equiformer: Equivariant Graph Attention Transformer for 3D Atomistic Graphs. International Conference on Learning Representations. 2023. [Google Scholar]
- Chen C., Ye W., Zuo Y., Zheng C., Ong S. P.. Graph Networks as a Universal Machine Learning Framework for Molecules and Crystals. Chem. Mater. 2019;31:3564–3572. doi: 10.1021/acs.chemmater.9b01294. [DOI] [Google Scholar]
- Wu Z., Jain P., Wright M., Mirhoseini A., Gonzalez J. E., Stoica I.. Representing Long-Range Context for Graph Neural Networks with Global Attention. arXiv. 2021:13266–13279. doi: 10.48550/arXiv.2201.08821. [DOI] [Google Scholar]
- Cheng J., Zhang C., Dong L.. A Geometric-Information-Enhanced Crystal Graph Network for Predicting Properties of Materials. Commun. Mater. 2021;2:1–11. doi: 10.1038/s43246-021-00194-3. [DOI] [Google Scholar]
- Yu, H. ; Hong, L. ; Chen, S. ; Gong, X. ; Xiang, H. . Capturing Long-Range Interaction with Reciprocal Space Neural Network. arXiv, Nov. 30, 2022. 10.48550/arXiv.2211.16684 [DOI]
- Batatia, I. ; Kovacs, D. P. ; Simm, G. N. C. ; Ortner, C. ; Csanyi, G. . MACE: Higher Order Equivariant Message Passing Neural Networks for Fast and Accurate Force Fields. arXiv. Jan. 26, 2022, ver. 2. 10.48550/arXiv.2206.07697 [DOI]
- Kovács D. P., Batatia I., Arany E. S., Csányi G.. Evaluation of the MACE Force Field Architecture: From Medicinal Chemistry to Materials Science. J. Chem. Phys. 2023;159:044118. doi: 10.1063/5.0155322. [DOI] [PubMed] [Google Scholar]
- Sadeghi A., Ghasemi S. A., Schaefer B., Mohr S., Lill M. A., Goedecker S.. Metrics for Measuring Distances in Configuration Spaces. J. Chem. Phys. 2013;139:184118. doi: 10.1063/1.4828704. [DOI] [PubMed] [Google Scholar]
- Zhu L., Amsler M., Fuhrer T., Schaefer B., Faraji S., Rostami S., Ghasemi S. A., Sadeghi A., Grauzinyte M., Wolverton C., Goedecker S.. A Fingerprint Based Metric for Measuring Similarities of Crystalline Structures. J. Chem. Phys. 2016;144:034203. doi: 10.1063/1.4940026. [DOI] [PubMed] [Google Scholar]
- Jain A., Ong S. P., Hautier G., Chen W., Richards W. D., Dacek S., Cholia S., Gunter D., Skinner D., Ceder G., Persson K. A.. Commentary: The Materials Project: A Materials Genome Approach to Accelerating Materials Innovation. APL Materials. 2013;1:011002. doi: 10.1063/1.4812323. [DOI] [Google Scholar]
- Dunn A., Wang Q., Ganose A., Dopp D., Jain A.. Benchmarking Materials Property Prediction Methods: The Matbench Test Set and Automatminer Reference Algorithm. npj Comput. Mater. 2020;6:1–10. doi: 10.1038/s41524-020-00433-0. [DOI] [Google Scholar]
- Kresse G., Furthmüller J.. Efficient Iterative Schemes for Ab Initio Total-Energy Calculations Using a Plane-Wave Basis Set. Phys. Rev. B. 1996;54:11169–11186. doi: 10.1103/PhysRevB.54.11169. [DOI] [PubMed] [Google Scholar]
- Kresse G., Joubert D.. From Ultrasoft Pseudopotentials to the Projector Augmented-Wave Method. Phys. Rev. B. 1999;59:1758–1775. doi: 10.1103/PhysRevB.59.1758. [DOI] [Google Scholar]
- Hjorth Larsen A., Jørgen Mortensen J., Blomqvist J., Castelli I. E, Christensen R., Dułak M., Friis J., Groves M. N, Hammer B.ør., Hargus C.. et al. The Atomic Simulation EnvironmentA Python Library for Working with Atoms. J. Phys.: Condens. Matter. 2017;29:273002. doi: 10.1088/1361-648X/aa680e. [DOI] [PubMed] [Google Scholar]
- Castelli I. E., Olsen T., Datta S., Landis D. D., Dahl S., Thygesen K. S., Jacobsen K. W.. Computational Screening of Perovskite Metal Oxides for Optimal Solar Light Capture. Energy Environ. Sci. 2012;5:5814–5819. doi: 10.1039/C1EE02717D. [DOI] [Google Scholar]
- Castelli I. E., Landis D. D., Thygesen K. S., Dahl S., Chorkendorff I., Jaramillo T. F., Jacobsen K. W.. New Cubic Perovskites for One- and Two-Photon Water Splitting Using the Computational Materials Repository. Energy Environ. Sci. 2012;5:9034–9043. doi: 10.1039/c2ee22341d. [DOI] [Google Scholar]
- Castelli I. E., Thygesen K. S., Jacobsen K. W.. Calculated Pourbaix Diagrams of Cubic Perovskites for Water Splitting: Stability Against Corrosion. Top Catal. 2014;57:265–272. doi: 10.1007/s11244-013-0181-4. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The source code and data for the EOSnet model are available at the following GitHub repository: https://github.com/Rutgers-ZRG/EosNet.


