Abstract
To precisely and reasonably describe the contribution of interatomic and intermolecular interactions to the physicochemical properties of complex systems, a chemical message passing strategy as driven by graph neural network is proposed. Thus, by distinguishing inherent and environmental features of atoms, as well as proper delivering of these messages upon growth of systems from atoms to bulk level, the evolution of system features affords eventually the target properties like the adsorption wavelength, emission wavelength, solubility, photoluminescence quantum yield, ionization energy, and lipophilicity. Considering that such a model combines chemical principles and natural behavior of atom aggregation crossing multiple scales, most likely, it will be proven to be rational and efficient for more general aims in dealing with complex systems.
Introduction
The rapid development of current chemical and materials industries requires rapid and accurate prediction of various properties involving multiple scales. However, the advances of quantum mechanics and classic thermodynamics cannot cover all these scales, which poses a challenge to the rational, ab initio design of higher-performance systems [1]. Thanks to the development of data-driven machine learning approaches [2,3], new energy has been infused into the fields of materials science and chemistry [2,4–10]. Thus, a sufficient number of samples can be used to provide a range of molecular properties through the trained model, which greatly accelerates the discovery of new functional molecules [11–14]. In predicting chemical properties, it is necessary to convert the chemical structure into computer-recognizable form of data. To this end, molecular descriptors [15] and molecular fingerprints [16] have been employed and proven to be useful for predicting various properties. However, these features may not be directly related to the properties of the molecule; they thus do not reflect the nature of interatomic or intermolecular interactions. Recently, machine learning models based on thermodynamic theory have also been used for prediction [17,18], but the complexity of feature processing limits generic application of these models. In any case, a fast and general model for multiscale property prediction is still lacking. Such a model is expected to be established based on chemical principles.
With the development of deep neural networks [19], representation learning has shown great advantages over feature based methods [20], attracting more and more attention concerning molecular property prediction. Graph neural network (GNN) learning for molecular representation has recently become an emerging research field that the topology of atoms and bonds is treated as a graph and propagates messages sent by each element to its neighbors [21,22]. However, although GNN can achieve feature transfer and aggregation from an atomic level, they still require guidance from certain chemical rules for the sake of efficiency and accuracy [23]. Further, in a specific system, the chemical environment is often more complex than the molecule itself. At this point, due to the inherent defects of the molecular graph, it is difficult to consider the chemical environment surrounding the molecule, such as solvent molecules, during message passing. Therefore, it remains challenging how to assign more chemical rules to the message passing process and to consider intermolecular interactions using GNN.
In order to endow GNN with more chemical rules, it is necessary to describe interatomic and intermolecular interactions reasonably. Based on this, more accurate message transmission is possible when subjecting a molecule in chemical environments. For this purpose, we have defined multiscale features (Fig. 1A), which are derived from undissolved molecules and transformed into 2 dissolved molecular graphs through different feature extensions (Fig. 1B). This allows information to gradually diffuse from intramolecular to intermolecular, fundamentally describing the contribution of each atom in the mixed system to the target properties (Fig. 1C). This natural growth neural network (NGNN) harmonizes chemical principles with algorithmic concepts, enabling accurate prediction of complex system properties.
Fig. 1.

Flow chart of the framework as developed in this work. (A) Multiscale features of 3 molecular graphs. (B) Features of 3 molecular graphs. (C) The working principle of NGNN.
Results and Discussion
GNN often used one-hot coding to encode the chemical environment of atoms, which whereas could not accurately describe the physical meaning of atoms. As shown in Fig. 2A, in order to regulate the model with more chemical rules, quite a few multiscale features have been introduced. For atomic features of specific molecules, they are classified into inherent atomic features () and atomic environment features (). The inherent features are determined by the element itself, while the environmental characteristics are exerted by each molecule. Furthermore, one can also label atoms or groups that have an important impact on the results (such as hydrophobic atoms for prediction of water solubility). Thus, from a chemical perspective, the formulation of these features are incompatible. Further, the inherent features chosen are all encoded in natural form, while the environmental features are encoded in one-hot form. Therefore, these features have to be mapped to separate spaces using different scales ( and ) to ensure that the extracted features can accurately describe the complete state of each atom in the molecule, and the calculation method is proposed as Eq. 1:
| (1) |
in which σ is activation function. For message passing between atoms, the edge features is used to drive the update of atomic nodes, using Eq. 2:
| (2) |
Fig. 2.

The workflow of NGNN. (A) Encoding and calculation of different node features. (B) Encoding and calculation of different edge features. (C) The way of message passing. (D) The way of multistage message passing between atoms and molecules and property prediction.
After node feature updates, the intact molecular graph is transformed into 2 forms of dissolved molecular graphs.
1. The description of solvent topology using electronic features, and the connectivity information and van der Waals radius to calculate the atomic contribution to a target property within the van der Waals surface area [24,25]. Starting from the charge and molecular structure of the solvent, the solvent effects are interpreted from dissolved molecules to quasi-dissolved atoms. Such dissolution of atoms proceeds via Eq. 3:
| (3) |
2. Applying the solvent effects on dissolved molecules to atomic interactions including bonding, long-range interactions, and concentration, that is, adding solvational interaction parameter [26,27] to the edge features. The solvent effects on each atom are considered in the course of information passing, as implemented using Eq. 4:
| (4) |
Next, message passing between molecules operates in 2 different ways (Eqs. 5 and 6):
| (5) |
| (6) |
After completing the message passing between molecules, the 2 dissolved molecular graphs are output via Eq. 7:
| (7) |
in which Set is a readout operator [28]. Finally, a separate neural network is used to predict molecular properties (Eq. 8:
| (8) |
To demonstrate the ability of the NGNN model to handle intermolecular interactions, 3 single-solvent datasets and 3 multisolvent datasets were employed for test with quite a few physicochemical properties as the target features, including the adsorption wavelength, emission wavelength, solubility, photoluminescence quantum yield (PLQY), ionization energy (IE), and lipophilicity (for more details, see Materials and Methods).
As shown in the previous text, we have designed a universal model using features that can essentially describe the inherent characteristics of each atom and the impact of chemical environment on the target properties. Through the transfer and aggregation of features between atoms, we ultimately obtain the molecular properties of specific environments, without the need to design specially targeted atomic and environmental features. On the other hand, for some functional groups or atoms sensitive to the target property, they can be marked out when describing the environmental characteristics. For example, the conjugated fragments can be labeled when predicting the spectroscopic properties, while the likely formed hydrogen bond should be considered when designing the features for predicting azeotropic points. As a more explicit instance, in predicting water solubility, we can label hydrophobic groups (such as F, Cl, Br, and I) in environmental features . As shown in Table 1, the labeled model exhibits higher accuracy than the unlabeled one.
Table 1.
The average performance of solubility in 3 independent runs
| Method | MAE | R 2 |
|---|---|---|
| NGNN-unmarked | 0.4410 | 0.8500 |
| NGNN-marked | 0.3887 | 0.8791 |
As shown in Fig. 3A and B, compared with other benchmark GNNs (graph convolutional network [GCN], graph attention network [GAT], and graph isomorphism network [GIN]) [29–31], NGNN exhibits high superiority. In the prediction of single-solvent data, the excellent performance of NGNN indicates that the interatomic message passing logic is indeed advanced in describing the interatomic and intermolecular interactions, in comparison with simple data stacking. Furthermore, in the prediction using multisolvent datasets, although with much smaller data volume and higher complexity (12 solvents), the superior performance of NGNN exhibits extremely high homogeneity for training set and prediction set (Fig. 3C).
Fig. 3.

Prediction of physicochemical properties with different models. (A) MAE of different models. (B) R2 of different models. (C) Training curves of different model in the prediction of absorption wavelength.
Next, the prediction of adsorption wavelength was employed to examine the robustness of different models. As shown in Fig. 4, NGNN is of stronger stability as compared to the previously reported models, and the errors are insignificant for both training and prediction. This indicates that NGNN can better understand the interactions within and between molecules and rationally apply them to unknown molecular predictions. Clearly, the message passing strategy proposed in this work can integrate the information of different solvent molecules with the characteristics of dissolved atoms and edges, thereby learning the influence of interconstituent interactions crossing multiple scales. Considering the higher dependency of absorption wavelength on both interatomic and intermolecular interactions, most likely, the NGNN model touches the essence of mixing.
Fig. 4.

The robustness of different models in the prediction of absorption wavelength. (A) Absolute error (AE) distribution of different models. (B) Absolute error distribution of different model training and test sets.
Further, to track the source of error, the error changes were traced all during the learning process. As shown in Fig. 5, the error changes for the train set evolve consistently with that for the test set in the modeling of absorption wavelength (Fig. 5B), which further proves the advance of the NGNN model in rationality. On the other hand, for water solubility prediction, due to the data available for only one solvent, the model relying solely on the properties of the solute itself, resulting in significant overfitting (Fig. 5A). Moreover, as shown in Table 2, traditional machine learning models (random forest [RF] and gradient boosting regression tree [GBRT]) [32,33] based on statistics lack guidance from chemical rules. To obtain better performance in the test, overfitting of the training set is unavoidable. In contrast, by considering the interatomic and intermolecular interactions from multiple perspectives, and through reasonable feature aggregation and propagation, NGNN is able to describe the characteristics of the system even upon scaling up from atomic level to bulk mixture. The macroscopic properties of a complex system can eventually be derived without obvious overfitting of the model.
Fig. 5.

Prediction results of solubility and absorption wavelength with different solvents using NGNN. (A) Training curves of NGNN in the prediction of solubility. (B) Training curves of NGNN in the prediction of absorption wavelength. (C) The distribution of MAE in the prediction of absorption wavelength. (D) Violin plot of MAE distribution under different solvent data in absorption spectrum prediction. (E) Prediction results for different solvents. (F) The correlation between the predicted and experimental values of the absorption wavelength.
Table 2.
The average performance of different models in predicting absorption wavelengths in 3 independent runs
| Training set | Test set | |||||||
|---|---|---|---|---|---|---|---|---|
| Model | MAE | RMSE | MRE | R 2 | MAE | RMSE | MRE | R 2 |
| GCN | 18.6570 | 27.1291 | 0.0420 | 0.9260 | 21.1936 | 35.4761 | 0.0477 | 0.8709 |
| GIN | 24.5175 | 33.6847 | 0.0557 | 0.8796 | 23.3730 | 37.0351 | 0.0523 | 0.8535 |
| GAT | 29.8693 | 43.9367 | 0.0675 | 0.7937 | 30.9669 | 46.0222 | 0.0691 | 0.7765 |
| NGNN | 10.7600 | 15.8400 | 0.0236 | 0.9800 | 10.6500 | 18.3300 | 0.0237 | 0.9600 |
| GBRT | 6.8296 | 9.8413 | 0.0160 | 0.9905 | 14.1696 | 25.0745 | 0.0324 | 0.9368 |
| RF | 4.8454 | 9.4383 | 0.0109 | 0.9912 | 13.0689 | 27.1370 | 0.0299 | 0.9258 |
To further demonstrate the behavior of NGNN, we separately modeled each solvent in the absorption wavelength dataset, and the results are shown in Fig. 5. There was a significant overfitting in the test set for each solvent prediction, which is consistent with the trend in the prediction of water solubility. On the contrary, in the case of unified modeling of all data, the errors of the training and testing sets show good consistency. Moreover, the mean absolute error (MAE) distribution (Fig. 5C to E) and correlation (Fig. 5F) of the training and testing sets further demonstrate the high generalization ability of NGNN. Due to previous data limitations (lack of data for solution properties of different concentrations), concentration was not taken into account in the feature. However, this feature can be considered by adding an term to modify the effect of solute–solvent ratio on the results. If the required data is available, most likely, the solution properties of different concentrations can be afforded.
Fig. 6.

Prediction results of 7 molecules with unprecedented elements used for testing. (A) MAE of NGNN model test set for comparison. (B) Prediction results containing Ge element molecules. (C) Prediction results containing Te element molecules. (D) Prediction results containing Se element molecules.
The design intention of our model is to extract the commonalities of each atomic feature as much as possible to gradually extend it to the entire molecule and complex environment. Thus, in order to better reflect the generalization of the model, we conducted extrapolation test on 7 external molecules, including 3 elements that do not exist in the training set. As shown in Fig. 6, the results showed that although our model have not learned molecules containing these elements, it is able to extract commonalities from the inherent features of atoms and expand them to unknown elements, which may greatly benefit the design of new molecules.
Conclusion
In summary, we have developed a NGNN framework for general, efficient prediction of a series of physicochemical properties based on proper handling of interatomic and intermolecular interactions of the system. The construction of this framework is based on strict chemical rules and natural behavior of system growing from atomic to bulk levels. By properly describing and transforming atomic features, message passing upon aggregating atoms to complex systems becomes more precise and efficient. Further, solvent features are embedded into molecular graphs in different ways through feature sharing, enabling GNN to handle complex intermolecular interactions. Such a framework is expected to compensate the shortcomings of GNN in predicting the properties of complex mixtures and may therefore be proven to be universally applicable in wider range of chemical/physical aims.
Materials and Methods
Data and processing
Water solubility dataset is a curated collection of the aqueous solubilities of organic compounds from 3 literature-based large databases: (a) Vermeire’s [34], (b) Boobier’s [35], and (c) Delaney’s [36]. The produced dataset was prepared by omitting the nonunique measures and noisy data, consisting of more than one solubility measure for a single molecule, yielding a total of 8,438 unique data entries [37]. Detailed information about IE and lipophilicity databases can be obtained on our GitHub website.
As shown in Tables 3 to 5, the absorption wavelength dataset, emission wavelength dataset, and PLQY dataset contain data for 12 solvents. The datasets are from the website (http://www.chemfluor.top).
Table 3.
The amount of data in the absorption wavelength database under different solvents
| Solvent | Toluene | THF | MeOH | MeCN | Hexane | H2O |
|---|---|---|---|---|---|---|
| Number | 254 | 243 | 314 | 289 | 98 | 623 |
| Solvent | EtOH | Ethyl acetate | DMSO | DMF | CHCl3 | CH2Cl2 |
| Number | 280 | 52 | 648 | 86 | 232 | 900 |
Table 5.
The amount of data under different solvents in the emission wavelength database
| Solvent | Toluene | THF | MeOH | MeCN | Hexane | H2O |
|---|---|---|---|---|---|---|
| Number | 268 | 243 | 314 | 289 | 98 | 623 |
| Solvent | EtOH | Ethyl acetate | DMSO | DMF | CHCl3 | CH2Cl2 |
| Number | 280 | 53 | 647 | 86 | 212 | 921 |
Table 4.
The amount of data under different solvents in the PLQY database
| Solvent | Toluene | THF | MeOH | MeCN | Hexane | H2O |
|---|---|---|---|---|---|---|
| Number | 232 | 200 | 203 | 258 | 80 | 219 |
| Solvent | EtOH | Ethyl acetate | DMSO | DMF | CHCl3 | CH2Cl2 |
| Number | 143 | 47 | 354 | 68 | 198 | 821 |
Method
Atomic inherent features ()
-
1.
Covalent radius.
-
2.
Electronegativity.
-
3.
Atomic number.
-
4.
Atomic mass.
-
5.
First IE.
-
6.
Electron affinity.
Note: These features are all encoded using natural coding.
Atomic environmental features ()
-
1.
Valence: (0, 1, 2, 3, 4, 5, 6).
-
2.
Number of H atoms: (0, 1, 2, 3, 4).
-
3.
Formal charge: (−1, −2, 1, 2, 0).
-
4.
Hybridization: (“s”, “sp”, “sp2”, “sp3”).
-
5.
Is inRing?: (1, 0).
-
6.
Is aromatic?: (1, 0).
-
7.
Custom features. Note: This feature can be defined based on the nature of the target and adopt one-hot coding.
Atomic environmental features (solvent) ()
-
1.
Electrotopological state (E-state) descriptors: This index combines the electronic states of intramolecular bonding atoms and their topological properties in the whole molecular skeleton. According to this descriptor, 3 internal states of the molecular substructure within the molecule are quantified: its element content, its valence state (electronic organization), and its topological state relative to its atomic neighbor [24].
-
2.
Molecular operating environment (MOE-type) descriptors: The MOE-type descriptors use connectivity information and van der Waals radii to calculate the atomic van der Waals surface area contribution of an atom-type to a given property, including polarizability, direct electrostatic interaction, and other factors [25].
-
3.
Topological descriptors: According to this descriptor, the connection state of each atom is used to calculate the exponent, thus providing a highly unique exponent for a given molecule [38].
-
4.
Connectivity descriptors.
Note: These features are all encoded using natural coding. This article utilizes 66 of the aforementioned features.
Interatomic features ()
-
1.
Bond type: (“single”, “double”, “triple”, “aromatic”).
-
2.
Conjugated: (1, 0).
-
3.
Aromatic: (1, 0).
-
4.
Ring: (1, 0).
Note: These features are all encoded using one-hot encoding.
Atomic features ()
-
1.
Et30 [26].
-
2.
SP: Polarizability [27].
-
3.
SdP: Dipolarity [27].
-
4.
SA: Acidity [27].
-
5.
SB: Basicity [27].
Note: This feature can be defined based on the nature of the target.
In each model training and prediction, we randomly select 80% of the data as the training set and the rest as the test set. We use the following indicators to evaluate the prediction performance of the model: mean relative error (MRE), MAE, root mean square error (RMSE), and the coefficient of determination (R2).
Acknowledgments
The authors thank the Zhejiang Xin’an Chemical Industrial Group Co. Ltd. for employing Prof. C. Qian as the specially appointed expert of the Qianjiang Talents Program in Hangzhou City.
Funding: Generous financial support was provided by the Key Research and Development Program of Zhejiang Province (2023C01102, 2023C01208, and 2022C01208).
Author contributions: J.F.: Methodology, investigation, data curation, and writing—original draft. C.Q.: Formal analysis and investigation. S.L.: Conceptualization, formal analysis, supervision, and writing—review.
Competing interests: The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this article.
Data Availability
The data used and more details (including the codes) about the framework are available at https://github.com/Fan1ing/ngnn.
References
- 1.Cao Y, Romero J, Olson JP, Degroote M, Johnson PD, Kieferova M, Kivlichan ID, Menke T, Peropadre B, Sawaya NPD, et al. Quantum chemistry in the age of quantum computing. Chem Rev. 2019;119(19):10856–10915. [DOI] [PubMed] [Google Scholar]
- 2.Butler KT, Davies DW, Cartwright H, Isayev O, Walsh A. Machine learning for molecular and materials science. Nature. 2018;559(7715):547–555. [DOI] [PubMed] [Google Scholar]
- 3.Granda JM, Donina L, Dragone V, Long DL, Cronin L. Controlling an organic synthesis robot with machine learning to search for new reactivity. Nature. 2018;559(7714):377–381. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Zhou MG, Liu ZP, Yin HL, Li CL, Xu TK, Chen ZB. Quantum neural network for quantum neural computing. Research. 2023;6:0134. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Zhang K, Nguyen T-T, Su Z, Demortière A. Self-supervised image quality assessment for X-ray tomographic images of Li-ion battery. npj Comput Mater. 2022;8(1):194. [Google Scholar]
- 6.Feng W, Wang L, Lin Z, Zhu Y, Wang H, Dong J, Bai R, Wang H, Zhou J, Peng W, et al. Generation of 3D molecules in pockets via a language model. Nat Mach Intell. 2024;6(1):62–73. [Google Scholar]
- 7.Guo S, Popp J, Bocklitz T. Chemometric analysis in Raman spectroscopy from experimental design to machine learning-based modeling. Nat Protoc. 2021;16(12):5426–5459. [DOI] [PubMed] [Google Scholar]
- 8.Soleimany AP, Amini A, Goldman S, Rus D, Bhatia SN, Coley CW. Evidential deep learning for guided molecular property prediction and discovery. ACS Cent Sci. 2021;7(8):1356–1367. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Ulissi ZW, Medford AJ, Bligaard T, Norskov JK. To address surface reaction network complexity using scaling relations machine learning and DFT calculations. Nat Commun. 2017;8:14621. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Li X, Zhu Q, Zhao C, Qian X, Zhang X, Duan X, Lin W. Tipping point detection using reservoir computing. Research. 2023;6:0174. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Fan J, Yuan B, Qian C, Zhou S. Group contribution method supervised neural network for precise design of organic nonlinear optical materials. Precis Chem. 2024;2(6):263–272. [Google Scholar]
- 12.Zhong M, Tran K, Min Y, Wang C, Wang Z, Dinh C-T, De Luna P, Yu Z, Rasouli AS, Brodersen P, et al. Accelerated discovery of CO2 electrocatalysts using active machine learning. Nature. 2020;581(7807):178–183. [DOI] [PubMed] [Google Scholar]
- 13.Gebauer NWA, Gastegger M, Hessmann SSP, Müller KR, Schütt KT. Inverse design of 3d molecular structures with conditional generative neural networks. Nat Commun. 2022;13(1):973. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Gómez-Bombarelli R, Wei JN, Duvenaud D, Hernández-Lobato JM, Sánchez-Lengeling B, Sheberla D, Aguilera-Iparraguirre J, Hirzel TD, Adams RP, Aspuru-Guzik A. Automatic chemical design using a data-driven continuous representation of molecules. ACS Cent Sci. 2018;4(2):268–276. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Carbonell P, Carlsson L, Faulon JL. Stereo signature molecular descriptor. J Chem Inf Model. 2013;53(4):887–897. [DOI] [PubMed] [Google Scholar]
- 16.Li S, Liu Y, Chen D, Jiang Y, Nie Z, Pan F. Encoding the atomic structure for machine learning in materials science. WIRES Comput Mol Sci. 2021;12(1):e1558. [Google Scholar]
- 17.Fan J, Qian C, Zhou S. Machine learning spectroscopy using a 2-stage, generalized constituent contribution protocol. Research. 2023;6:0115. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Yu L-Y, Ren G-P, Hou X-J, Wu K-J, He Y. Transition state theory-inspired neural network for estimating the viscosity of deep eutectic solvents. ACS Cent Sci. 2022;8(7):983–995. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521(7553):436–444. [DOI] [PubMed] [Google Scholar]
- 20.Fang X, Liu L, Lei J, He D, Zhang S, Zhou J, Wang F, Wu H, Wang H. Geometry-enhanced molecular representation learning for property prediction. Nat Mach Intell. 2022;4(2):127–134. [Google Scholar]
- 21.Fang Y, Zhang Q, Zhang N, Chen Z, Zhuang X, Shao X, Fan X, Chen H. Knowledge graph-enhanced molecular contrastive learning with functional prompt. Nat Mach Intell. 2023;5(5):542–553. [Google Scholar]
- 22.Young A, Röst H, Wang B. Tandem mass spectrum prediction for small molecules using graph transformers. Nat Mach Intell. 2024;6(4):404–416. [Google Scholar]
- 23.Singh K, Munchmeyer J, Weber L, Leser U, Bande A. Graph neural networks for learning molecular excitation spectra. J Chem Theory Comput. 2022;18(7):4408–4417. [DOI] [PubMed] [Google Scholar]
- 24.Hall LH, Kier LB. Electrotopological state indices for atom types: A novel combination of electronic, topological, and valence state information. J Chem Inf Model. 1995;35(6):1039–1045. [Google Scholar]
- 25.Labute PA. Widely applicable set of descriptors. J Mol Graph Model. 2000;18(4–5):464–477. [DOI] [PubMed] [Google Scholar]
- 26.Reichardt C. Solvatochromic dyes as solvent polarity indicators. Chem Rev. 1994;94(8):2319–2358. [Google Scholar]
- 27.Catalán J. Toward a generalized treatment of the solvent effect based on four empirical scales: Dipolarity (SdP, a new scale), polarizability (SP), acidity (SA), and basicity (SB) of the medium. J Phys Chem B. 2009;113(17):5951–5960. [DOI] [PubMed] [Google Scholar]
- 28.Oriol V, Samy B, Manjunath K. Order matters: Sequence to sequence for sets. arXiv. 2015. 10.48550/arXiv.1511.06391 [DOI]
- 29.Kipf TN, Welling M. Semi-supervised classification with graph convolutional networks. In Proc. 5th International Conference on Learning Representations (OpenReview.net, 2017).
- 30.Xu K, Hu W, Leskovec J, Jegelka S. How powerful are graph neural networks? In Proc. 7th International Conference on Learning Representations (OpenReview.net, 2019).
- 31.Brody S, Alon U, Yahav E. How attentive are graph attention networks? arXiv. 2022. 10.48550/arXiv.2105.14491 [DOI]
- 32.Breiman L. Random forests. Mach Learn. 2001;45:5–32. [Google Scholar]
- 33.Friedman JH. Greedy function approximation: A gradient boosting machine. Ann Stat. 2001;29:1189–1232. [Google Scholar]
- 34.Vermeire FH, Chung Y, Green WH. Predicting solubility limits of organic solutes for a wide range of solvents and temperatures. J Am Chem Soc. 2022;144(24):10785–10797. [DOI] [PubMed] [Google Scholar]
- 35.Boobier S, Hose DRJ, Blacker AJ, Nguyen BN. Machine learning with physicochemical relationships: Solubility prediction in organic solvents and water. Nat Commun. 2020;11(1):5753. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Delaney JS. ESOL: Estimating aqueous solubility directly from molecular structure. J Chem Inf Comput Sci. 2004;44(3):1000–1005. [DOI] [PubMed] [Google Scholar]
- 37.Tayyebi A, Alshami AS, Rabiei Z, Yu X, Ismail N, Talukder MJ, Power J. Prediction of organic compound aqueous solubility using machine learning: A comparison study of descriptor-based and fingerprints-based models. J Cheminform. 2023;15(1):99. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Narumi H. New topological indices for finite and infinite systems. MATCH Commun Math Comput Chem. 1987;22:195–207. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data used and more details (including the codes) about the framework are available at https://github.com/Fan1ing/ngnn.
