Skip to main content
PLOS One logoLink to PLOS One
. 2025 Apr 2;20(4):e0319917. doi: 10.1371/journal.pone.0319917

Assessing interface accuracy in macromolecular complexes

Olgierd Ludwiczak 1, Maciej Antczak 1,2,*, Marta Szachniuk 1,2,*
Editor: Yong Wang3
PMCID: PMC11964455  PMID: 40173387

Abstract

Accurately predicting the 3D structures of macromolecular complexes is becoming increasingly important for understanding their cellular functions. At the same time, reliably assessing prediction quality remains a significant challenge in bioinformatics. To address this, various methods analyze and evaluate in silico models from multiple perspectives, accounting for both the reconstructed components’ structures and their arrangement within the complex. In this work, we introduce Intermolecular Interaction Network Fidelity (I-INF), a normalized similarity measure that quantifies intermolecular interactions in multichain complexes. Adapted from a well-established score in the RNA field, I-INF provides a clear and intuitive way to evaluate the predicted 3D models against a reference structure, with a specific focus on interchain interaction sites. Additionally, we implement the F1 measure to assess interfaces in macromolecular assemblies, further enriching the evaluation framework. Tested on 72 RNA-protein decoys, as well as exemplary DNA-DNA, RNA-RNA, and protein-protein complexes, these measures deliver reliable scores and enable straightforward ranking of predictions. The tool for computing I-INF and F1 is publicly available on Zenodo, facilitating large-scale analysis and integration with other computational systems.

Introduction

Molecular complexes play crucial roles in various cellular processes, including gene expression and homeostasis. Understanding their biological functions relies on detailed structural studies that reveal the conformations of constituent molecules and the mechanisms by which they form stable complexes. Traditionally, such studies have employed experimental techniques, however, in recent years, computational prediction methods have gained prominence, generating increasingly accurate and reliable models. These methods undergo systematic evaluation in blind prediction challenges, such as CASP or RNA-Puzzles, where computational predictions are assessed both in the context of reference structures and independently from stereochemistry and chain topology perspectives [15].

Evaluating molecular complex predictions is inherently challenging and often involves a variety of scoring functions, including knowledge- and machine-learning-based approaches [6]. Some of these functions were initially developed for isolated proteins or nucleic acids and later adapted to assess their complexes. For example, RMSD was adapted to score the interface between various chains of predicted multimeric assemblies and was applied as I-RMSD (Interface RMSD) to RNA-ligand predictions [7]. Along with LRMSD (Ligand RMSD), I-RMSD has become part of DockQv2 [8], which is aimed at evaluating the accuracy of complexes involving proteins, nucleic acids, and small molecules. Other scoring functions developed specifically for multimers include oligolDDT [9] and US-align [10]. In general, the evaluation of molecular complexes proceeds in two ways: by assessing the quality of interactions between chains or overall structural similarity.

Intermolecular Interaction Network Fidelity (I-INF), introduced here to assess the prediction of macromolecular complexes, is an adaptation of the RNA-specific INF score [11], with the focus shifting from base pairs to intermolecular interactions. Like the original, I-INF is a similarity measure ranging from 0 to 1, where 0 indicates a completely incorrect prediction and 1 signifies that the prediction is fully consistent with the reference structure. Tested on 72 structures from RNA-protein docking decoys [13], I-INF shows a high correlation with TM-score-based rankings [14] and a low correlation with DockQv2 [8]. This complementary nature highlights I-INF’s usefulness as part of a broader toolkit for evaluating three-dimensional macromolecular models. To provide additional flexibility for users, we also implement the F1 score, which is widely used for evaluating protein structure predictions, particularly focusing on hydrogen bonds [12]. While both I-INF and F1 assess the same aspect of macromolecular assemblies, their mathematical formulations differ: I-INF uses a geometric mean, whereas F1 employs a harmonic mean. Similar to I-INF, we adapt F1 to evaluate interchain interactions; in both measures, an interaction is counted as a true positive regardless of the number of hydrogen bonds in the predicted model. Together, I-INF and F1 provide a robust framework for assessing intermolecular interfaces in macromolecular complexes.

Materials and methods

Data processing for computing I-INF and F1 involves three steps: preprocessing the 3D structure data, quantifying hydrogen bonds that form intermolecular interactions in both the predicted and native structures, and rescaling the score based on the target coverage by predicted residues (Fig 1).

Fig 1. Data flow in the assessment of macromolecular assembly predictions.

Fig 1

In the input stage, users provide the reference and predicted 3D structure(s) in PDB format. They may also ensure consistent residue mapping between each model and the reference structure as well as provide additional information about irrelevant residues and a scaling flag. Preprocessing begins by filtering out irrelevant residues in the input structures, if specified. Next, the 3D structure data are processed using rna-tools [15] to align with the provided residue mapping, if available. In the third phase, HBPLUS [16] is executed to identify hydrogen bonds between RNA, DNA, and protein chains in each molecular assembly. If residue mapping between the predicted model and the target is not predefined, an additional algorithm is employed to determine all maximum mappings. This algorithm operates on a bipartite graph that represents the sequences of the analyzed structures and identifies the maximum bipartite matching within the graph [17]. Subsequently, pairs of binding residues are extracted from both the target and predicted models regardless of the number of hydrogen bonds they form. Each residue pair is categorized as a true positive (TP; present in both the target and prediction), false positive (FP; present only in the prediction), or false negative (FN; present only in the target). Finally, the I-INF and F1 scores are computed for the predicted model:

IINF=|TP||TP|+|FP||TP||TP|+|FN|F1=|TP||TP|+12(|FP|+|FN|)

After calculating the scores, if multiple mappings exist for a predicted structure, only the one with the highest I-INF (and F1) score is selected for the subsequent steps. Postprocessing involves the optional rescaling of the I-INF and F1 values by multiplying them by the fraction of predicted residues contained in the target. For example, if a target has 100 residues and 80 of those are predicted in the model, the scale factor would be 0.8. The predicted models are then sorted in non-increasing order by I-INF, and a list of model names with their assigned I-INF and F1 values is output in a CSV file.

The I-INF tool was developed in Python 3 and is available under the MIT license, with ready-to-run examples. It is published on GitHub and Zenodo, and supports the processing of various types of intermolecular complexes (e.g., RNA-RNA, DNA-DNA, protein- protein).

Results and discussion

To test and validate I-INF, we applied it to evaluate RNA-protein docking decoys, consisting of 72 experimental 3D structures of varying complexity along with their in silico generated models. Fig 2 shows a sample model-target pair from this collection, highlighting the intermolecular interactions. In the target structure (PDB ID: 3MOJ [18]), there are 4 interactions forming hydrogen bonds, whereas, in the predicted model, there are 9, of which 4 are true positives and the remaining 5 are considered false positives. In this case, I-INF is 0.67.

Fig 2. (A) Reference structure (PDB ID: 3MOJ) and (B) the predicted model of the RNA binding domain of the Bacillus subtilis YxiN protein complexed with a fragment of 23S ribosomal RNA. Residues involved in RNA-protein binding are color-coded: green for true positives (interactions present in both the reference structure and the model), red for false positives (interactions present only in the predicted model), and orange for 3 residues that form a multiplet in the predicted model, where one interaction is a true positive and the other is a false positive. No false negatives (interactions present only in the reference structure) are observed for this pair of structures.

Fig 2

For comparison, we evaluated all models from the benchmark set using two other methods, TM-score [14] and DockQv2 [8]. Both are normalized similarity measures that take values between 0 and 1. The numerical results of this comparative analysis are provided in S1 Table. Fig 3 visualizes the correlations between TM-score, DockQv2, and I-INF, while S1 Fig illustrates the distribution of these measures across the benchmark set. We also analyzed the Pearson correlation between the score rankings generated by these three measures. Although TM-score evaluates the global topology of the model and I-INF specifically assesses the accuracy of the intermolecular interface, we observed a high Pearson correlation (0.73) between these metrics. This suggests that, in our dataset, a correctly predicted global fold is largely a consequence of the proper spatial arrangement of the molecular components, resulting in accurately modeled interfaces. In other words, deviations in the overall fold are primarily associated with errors in the interface regions. Therefore, high TM-score values frequently coincide with high I-INF values, highlighting our models’ interdependence between global structural accuracy and interface correctness. In contrast, the low correlation with DockQv2 (0.18) indicates that it provides different insights than I-INF. Thus, these two measures are complementary, and it is beneficial to use both in evaluations.

Fig 3. Correlation between TM-score and I-INF (left) and DockQv2 and I-INF (right).

Fig 3

In an additional computational experiment, we analyzed the results of two RNA-protein targets from CASP15. The native assemblies consisted of one RNA strand and six protein chains for RT1189 (PDB ID: 7YR7 [19]) and one RNA strand and four protein chains for RT1190 (PDB ID: 7YR6 [19]). While the predicted complexes often included accurate protein or RNA structures, their chains were not properly docked, as indicated by low TM-score values (none of the models exceeded 0.5). This was further confirmed by I-INF calculations, which were close to zero in all cases.

Conclusions

In this study, we presented the I-INF metric (Intermolecular Interaction Network Fidelity) as a novel approach for evaluating the accuracy of intermolecular interactions within macromolecular complexes. I-INF provides a complementary measure to existing scoring functions, such as TM-score and DockQv2, offering a robust tool for assessing the interfaces in RNA-protein and other macromolecular complexes. Additionally, we have incorporated the F1 as part of our evaluation process, allowing for a more comprehensive comparison with other commonly used measures in structural prediction.

The tool for calculating I-INF and F1 operates on input PDB files, which is currently the only supported format due to limitations of HBPLUS. In future updates, we plan to expand its functionality by adding support for the mmCIF format, which will enhance compatibility with other widely used structural databases and prediction tools. As the quality of predictions improves, we may increase the sensitivity of our approach by focusing on hydrogen bonding interactions, rather than only on residue pairs involved in these interactions.

Supporting information

S1 Table. Evaluation of the predicted 3D models from the RNA-protein docking decoys.

(PDF)

pone.0319917.s001.pdf (265.2KB, pdf)
S1 Fig. A distribution of TM-score, DockQv2, and I-INF values computed for the benchmark set.

(PDF)

pone.0319917.s002.pdf (91.7KB, pdf)

Acknowledgments

This work was carried out at Poznan University of Technology (https://www.put.poznan.pl/en) and the Institute of Bioorganic Chemistry, Polish Academy of Sciences (https://www.ibch.poznan.pl/en.html). The authors are grateful for the support and resources provided by the institution.

Data Availability

The tool for computing I-INF and F1 to assess interfaces in macromolecular assemblies, along with a user manual and ready-to-run examples, is publicly available at Zenodo (https://dx.doi.org/10.5281/ zenodo.14697284) and GitHub (https://github.com/OlgierdL/iinf). For benchmarking, we used the dataset available at https://zoulab.dalton.missouri.edu/RNAdecoys/index.html.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Wodak SJ, Vajda S, Lensink MF, Kozakov D, Bates PA. Critical assessment of methods for predicting the 3D structure of proteins and protein complexes. Annu Rev Biophys. 2023;52183–206. doi: 10.1146/annurev-biophys-102622-084607 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Das R, Kretsch RC, Simpkin AJ, Mulvaney T, Pham P, Rangan R, et al. Assessment of three-dimensional RNA structure prediction in CASP15. Proteins 2023;91(12):1747–70. doi: 10.1002/prot.26602 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Carrascoza F, Antczak M, Miao Z, Westhof E, Szachniuk M. Evaluation of the stereochemical quality of predicted RNA 3D models in the RNA-Puzzles submissions. RNA 2022;28(2):250–62. doi: 10.1261/rna.078685.121 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Popenda M, Zok T, Sarzynska J, Korpeta A, Adamiak RW, Antczak M, et al. Entanglements of structure elements revealed in RNA 3D models. Nucleic Acids Res 2021;49(17):9625–32. doi: 10.1093/nar/gkab716 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Gren BA, Antczak M, Zok T, Sulkowska JI, Szachniuk M. Knotted artifacts in predicted 3D RNA structures. PLoS Comput Biol 2024;20(6):e1011959. doi: 10.1371/journal.pcbi.1011959 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Zeng C, Zhuo C, Gao J, Liu H, Zhao Y. Advances and challenges in scoring functions for RNA-protein complex structure prediction. Biomolecules 2024;14(10):1245. doi: 10.3390/biom14101245 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Nithin C, Kmiecik S, Błaszczyk R, Nowicka J, Tuszyńska I. Comparative analysis of RNA 3D structure prediction methods: towards enhanced modeling of RNA-ligand interactions. Nucleic Acids Res 2024;52(13):7465–86. doi: 10.1093/nar/gkae541 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Mirabello C, Wallner B. DockQv2: Improved automatic quality measure for protein multimers, nucleic acids and small molecules. Bioinformatics. 2024;40(10):btae586. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Haas J, Gumienny R, Barbato A, Ackermann F, Tauriello G, Bertoni M, et al. Introducing “best single template’’ models as reference baseline for the Continuous Automated Model Evaluation (CAMEO). Proteins 2019;87(12):1378–87. doi: 10.1002/prot.25815 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Zhang C, Shine M, Pyle AM, Zhang Y. US-align: universal structure alignments of proteins, nucleic acids, and macromolecular complexes. Nat Methods 2022;19(9):1109–15. doi: 10.1038/s41592-022-01585-1 [DOI] [PubMed] [Google Scholar]
  • 11.Parisien M, Cruz JA, Westhof E, Major F. New metrics for comparing and assessing discrepancies between RNA 3D structures and models. RNA 2009;15(10):1875–85. doi: 10.1261/rna.1700409 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Nugent T, Cozzetto D, Jones DT. Evaluation of predictions in the CASP10 model refinement category. Proteins. 2014;82(Suppl 2):98–111. doi: 10.1002/prot.24377 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Huang S-Y, Zou X. A nonredundant structure dataset for benchmarking protein-RNA computational docking. J Comput Chem 2013;34(4):311–8. doi: 10.1002/jcc.23149 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Gong S, Zhang C, Zhang Y. RNA-align: quick and accurate alignment of RNA 3D structures based on size-independent TM-scoreRNA. Bioinformatics 2019;35(21):4459–61. doi: 10.1093/bioinformatics/btz282 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Magnus M, Antczak M, Zok T, Wiedemann J, Lukasiak P, Cao Y, et al. RNA-Puzzles toolkit: a computational resource of RNA 3D structure benchmark datasets, structure manipulation, and evaluation tools. Nucleic Acids Res 2020;48(2):576–88. doi: 10.1093/nar/gkz1108 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.McDonald IK, Thornton JM. Satisfying hydrogen bonding potential in proteins. J Mol Biol. 1994;5(238):777–93. [DOI] [PubMed] [Google Scholar]
  • 17.Uno T. Algorithms for enumerating all perfect, maximum and maximal matchings in bipartite graphs. Algorithms and Computation: 8th International Symposium, ISAAC’97 Singapore, 1997 December 17–19. vol. 8. 1997. p. 92–101. [Google Scholar]
  • 18.Hardin JW, Hu YX, McKay DB. Structure of the RNA binding domain of a DEAD-box helicase bound to its ribosomal RNA target reveals a novel mode of recognition by an RNA recognition motif. J Mol Biol 2010;402(2):412–27. doi: 10.1016/j.jmb.2010.07.040 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Jia X, Pan Z, Yuan Y, Luo B, Luo Y, Mukherjee S, et al. Structural basis of sRNA RsmZ regulation of Pseudomonas aeruginosa virulence. Cell Res 2023;33(4):328–30. doi: 10.1038/s41422-023-00786-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
PLoS One. 2025 Apr 2;20(4):e0319917. doi: 10.1371/journal.pone.0319917.r001

Author response to Decision Letter 0


Transfer Alert

This paper was transferred from another journal. As a result, its full editorial history (including decision letters, peer reviews and author responses) may not be present.

20 Nov 2024

Decision Letter 0

Yong Wang

15 Dec 2024

PONE-D-24-53421The I-INF metric for assessing the quality of macromolecular assembly predictionPLOS ONE

Dear Dr. Szachniuk,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

==============================

The reviews highlight both the potential utility of I-INF and several significant concerns regarding its novelty, theoretical foundation, and applicability. While one reviewer expressed overall satisfaction, their sole recommendation—a comparative figure—should be considered alongside the more extensive issues raised by the other reviewer. The manuscript provides a valuable contribution to the evaluation of structural prediction methods. However, addressing the outlined deficiencies is essential to ensure the robustness, usability, and contextual grounding of I-INF as a broadly applicable metric.

==============================

Please submit your revised manuscript by Jan 29 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Yong Wang

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, we expect all author-generated code to be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. Thank you for uploading your study's underlying data set. Unfortunately, the repository you have noted in your Data Availability statement does not qualify as an acceptable data repository according to PLOS's standards. At this time, please upload the minimal data set necessary to replicate your study's findings to a stable, public repository (such as figshare or Dryad) and provide us with the relevant URLs, DOIs, or accession numbers that may be used to access these data. For a list of recommended repositories and additional information on PLOS standards for data deposition, please see https://journals.plos.org/plosone/s/recommended-repositories.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This manuscript describes I-INF, a metric to evaluate the correctness of inter-molecular hydrogen bonding. Both the code and the manuscript is overly simple. Major issues include:

1. Hydrogen bonding evaluation is routinely done in CASP, e.g., in https://pmc.ncbi.nlm.nih.gov/articles/PMC4282348/, except that I-INF takes the geometric average while CASP uses harmonic average. Please properly credit earlier works and explain the differences.

2. TM-score an I-INF does NOT measure the same aspect of complex assembly. I-INF only evaluate the correctness of interface, while TM-score evaluates the global topology. There should be some theoretical justification on why TM-score has a high correlation with I-INF.

3. The title strongly suggest that I-INF works even for any bio-macromolecular complexes including pure protein complexes. However, all evaluations in the result section are on protein-RNA complexes. There should be some application of I-INF on protein-protein complexes or explanation on why this is not applicable.

4. A major challenge in complex assembly evaluation is symmetric. For example, suppose the input is a tetramer containing two identical copies of a protein and two identical copies of an RNA, where the two RNA chains interact with each other. In this case, deriving a correct chain mapping for optimal I-INF is not trivial. This work simply ignore this challenge by forcing the user to input consistent chain ID mapping. This makes the program less useful.

5. There is no support for mmCIF format, which is particularly relevant for this work which is supposed to work with large complexes with >65 chains.

Reviewer #2: This work studies the evaluation of 3D structural prediction for macromolecular complexes, which is an essential problem in understanding cellular functions. The authors propose a novel metric, I-INF (Intermolecular Interaction Network Fidelity), to quantitatively assess the intermolecular interactions within multichain complexes. This manuscript is very well written and the I-INF score is useful. And I have only one minor comment about it.

When comparing I-INF with existing methods such as TM-score and DockQv2, it is recommended to add a figure to make the comparison more clear.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2025 Apr 2;20(4):e0319917. doi: 10.1371/journal.pone.0319917.r003

Author response to Decision Letter 1


29 Jan 2025

We thank the Reviewers for all the comments and suggestions. We responded to them to the best of our ability and made the required changes to the manuscript.

In line with PLOS ONE recommendations, we integrated the Zenodo platform with our tool repository on GitHub, enabling us to assign a unique DOI for each new release. The current DOI for the I-INF tool (version 1.0) is 10.5281/zenodo.14697284, and it can be accessed via https://zenodo.org/records/14697284 or https://dx.doi.org/10.5281/zenodo.14697284. All future updates will also be made available through these links.

Reviewer 1

This manuscript describes I-INF, a metric to evaluate the correctness of inter-molecular hydrogen bonding. Both the code and the manuscript is overly simple. Major issues include:

{Comment 1:} Hydrogen bonding evaluation is routinely done in CASP, e.g., in https://pmc.ncbi.nlm.nih.gov/articles/PMC4282348/, except that I-INF takes the geometric average while CASP uses harmonic average. Please properly credit earlier works and explain the differences.

{Response:} We thank the Reviewer for highlighting the evaluation approach used in CASP, particularly the application of the harmonic mean (e.g., F1 score) for assessing hydrogen bond predictions. While CASP assessors commonly employ the harmonic mean (F1 score), the RNA-Puzzles contest has predominantly utilized the geometric mean (INF score) for evaluating RNA-related predictions. A key distinction between the evaluation methods in CASP and RNA-Puzzles lies in the granularity of interaction assessment. In CASP, individual hydrogen bonds are explicitly evaluated when calculating the F1 score. In contrast, RNA-Puzzles focuses on broader interaction assessments rather than individual hydrogen bonds. For instance, if two nucleotides in the reference structure are connected by three hydrogen bonds, but the predicted model shows the same nucleotides forming only two hydrogen bonds, the interaction is still considered a true positive. This approach disregards discrepancies in the exact number of hydrogen bonds, focusing instead on the presence of the interaction itself rather than its precise details. Such a solution seems more suitable for predictions that have not yet reached exceptional accuracy.

In response to the Reviewer’s comment, we have implemented the F1 score in our scripts to enable direct comparison with I-INF. In both I-INF and F1 measures, an interaction is counted as a true positive regardless of the number of hydrogen bonds in the predicted model. Consequently, we updated Figure 1 (depicting the workflow) and revised the description of our program for molecular complex evaluation to reflect this addition. We also updated the manuscript title to emphasize the inclusion of multiple evaluation methods and our commitment to presenting a balanced and comprehensive analysis.

Finally, it is worth noting that when evaluating multiple predictions, our program ranks them based on the value of the evaluation function. We designated I-INF as the primary arbiter, as substituting it with F1 would not affect the final ranking.

{Comment 2:} TM-score an I-INF does NOT measure the same aspect of complex assembly. I-INF only evaluate the correctness of interface, while TM-score evaluates the global topology. There should be some theoretical justification on why TM-score has a high correlation with I-INF.

{Response:} We appreciate the comment of the Reviewer on the correlation between TM-score and I-INF. In our opinion, the justification for the observed high correlation between these two measures is closely tied to the characteristics of the validation dataset. In general, the tertiary structures of the models closely resemble the reference structures, with the primary differences occurring at the interface regions, which is the focus of our evaluation.

The validation set was specifically designed to assess the quality of methods evaluating intermolecular interfaces. Since I-INF is focused on assessing the accurracy of the interfaces, while TM-score evaluates the global topology, the significant overlap in terms of structural correctness at the interface region likely accounts for the high correlation between the two measures. This makes sense, as both metrics are sensitive to the accuracy of structural predictions, albeit with different scopes of evaluation.

{Comment 3:} The title strongly suggest that I-INF works even for any biomacromolecular complexes including pure protein complexes. However, all evaluations in the result section are on protein-RNA complexes. There should be some application of I-INF on protein-protein complexes or explanation on why this is not applicable.

{Response:} We thank the Reviewer for this observation. First, as mentioned in our response to Comment 1, we have updated the manuscript title to remove the explicit reference to I-INF. This change ensures that the title no longer implies an exclusive focus on this measure.

Second, we agree that I-INF is not inherently limited to RNA-protein complexes. The measure is designed to evaluate the interface accuracy in any type of macromolecular complex. To demonstrate this, we have now included additional examples that apply I-INF to DNA-DNA, RNA-RNA, and protein-protein assemblies. They are available in our Zenodo repository (DOI: 10.5281/zenodo.14697284). This addition underscores the versatility of the I-INF measure and further substantiates its broader applicability beyond RNA-related systems.

{Comment 4:} A major challenge in complex assembly evaluation is symmetric. For example, suppose the input is a tetramer containing two identical copies of a protein and two identical copies of an RNA, where the two RNA chains interact with each other. In this case, deriving a correct chain mapping for optimal I-INF is not trivial. This work simply ignore this challenge by forcing the user to input consistent chain ID mapping. This makes the program less useful.

{Response:} We thank the Reviewer for raising this important point regarding the challenge of symmetric complexes. We acknowledge that deriving a correct chain mapping for complex assemblies, especially in cases of identical components such as in the example of tetramers, can be non-trivial. To address this issue, we have implemented a mechanism similar to that used in DockQv2. Specifically, we introduced an algorithm that efficiently iterates over all possible combinations of chains that can occur between the model and the reference structure. For each model, we now select the highest I-INF value from all analyzed chain mappings.

To achieve this, we construct a bipartite graph that includes two independent layers of vertices corresponding to chains in the model and the reference structure. An edge is created between two vertices if the corresponding chains have identical sequences. We analyze all possible maximum matchings in this graph using Uno's algorithm (1997), which enumerates perfect, maximum, and maximal matchings in bipartite graphs.

For the implementation, we used open-source code, which we refined by limiting it to the essential functions needed for our analysis. The source implementation is available at https://github.com/Xunius/bipartite_matching.

This approach enhances our tool's ability to handle complex assemblies with symmetric components, making it more versatile and applicable to a broader range of systems. Examples of RNA-RNA and protein-protein interactions are provided (in our Zenodo repository) to illustrate its application.

{Comment 5:} There is no support for mmCIF format, which is particularly relevant for this work which is supposed to work with large complexes with $>$65 chains.

{Response:} Thank you for your feedback on the support for the mmCIF format. We acknowledge that the lack of mmCIF format support is a limitation, especially for large complexes with more than 65 chains, as you mentioned. As noted in the manuscript, the current version of the I-INF tool operates on PDB files due to the limitations of the HBPLUS library. However, we have already planned to implement support for the mmCIF format in future updates to address this limitation and improve the tool's applicability for larger RNA-protein complexes. We appreciate your valuable suggestion, and we will continue to work toward improving the capabilities of the tool in future versions.

Reviewer 2

This work studies the evaluation of 3D structural prediction for macromolecular complexes, which is an essential problem in understanding cellular functions. The authors propose a novel metric, I-INF (Intermolecular Interaction Network Fidelity), to quantitatively assess the intermolecular interactions within multichain complexes. This manuscript is very well written and the I-INF score is useful. And I have only one minor comment about it. When comparing I-INF with existing methods such as TM-score and DockQv2, it is recommended to add a figure to make the comparison more clear.

{Response:} We thank the Reviewer for the positive feedback and helpful suggestions. In response, we have added two scatter plots (Figure 3) to the manuscript that visualize the correlations between TM-score, DockQv2, and I-INF. Additionally, we have included a violin plot in the Supplementary Information (S1\_Figure.pdf) to illustrate how the values of these measures are spread across the benchmark set. Each violin corresponds to one measure, highlighting its central tendency and variability. We believe these additions enhance the clarity of our evaluation and strengthen the presentation of our findings.

Attachment

Submitted filename: Response-to-reviewers.pdf

pone.0319917.s003.pdf (89.1KB, pdf)

Decision Letter 1

Yong Wang

7 Feb 2025

PONE-D-24-53421R1Assessing interface accuracy in macromolecular complexesPLOS ONE

Dear Dr. Szachniuk,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

==============================

Thank you for your revised manuscript and your response to the reviewer’s comments. However, the reviewer has pointed out that while their concern regarding the correlation between TM-score and I-INF was addressed in the response letter, no corresponding modifications were made in the manuscript itself.

You should incorporate a theoretical justification within the manuscript explaining why TM-score has a high correlation with I-INF. This addition should ensure that the manuscript provides sufficient context and rationale for the use of these metrics in evaluating complex assembly.

Please revise the manuscript accordingly and highlight the changes in your response letter.

==============================

Please submit your revised manuscript by Mar 24 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Yong Wang

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I previously commented that "TM-score an I-INF does NOT measure the same aspect of complex assembly. I-INF only evaluate the correctness of interface, while TM-score evaluates the global topology. There should be some theoretical justification on why TM-score has a high correlation with I-INF." This revision only answer my question in the response letter, but did not make any corresponding modification in the manuscript.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2025 Apr 2;20(4):e0319917. doi: 10.1371/journal.pone.0319917.r005

Author response to Decision Letter 2


7 Feb 2025

We thank the Reviewer for the supplementary suggestion. We have answered it in the revised version of the manuscript, and below we provide detailed explanations:

Reviewer's comment:

I previously commented that 'TM-score and I-INF do NOT measure the same aspect of complex assembly. I-INF only evaluates the correctness of the interface, while TM-score evaluates the global topology. There should be some theoretical justification on why TM-score has a high correlation with I-INF.' This revision only answered my question in the response letter, but did not make any corresponding modification in the manuscript.

Response:

Indeed, in the previous version of the manuscript, we did not include the appropriate modification in response to this comment. We have now addressed this. In the "Results and Discussion" section, we have inserted the following text:

“Although TM-score evaluates the global topology of the model and I-INF specifically assesses the accuracy of the intermolecular interface, we observed a high Pearson correlation (0.73) between these metrics. This suggests that, in our dataset, a correctly predicted global fold is largely a consequence of the proper spatial arrangement of the molecular components, resulting in accurately modeled interfaces. In other words, deviations in the overall fold are primarily associated with errors in the interface regions. Therefore, high TM-score values frequently coincide with high I-INF values, highlighting our models' interdependence between global structural accuracy and interface correctness.”

We hope that this addition provides the necessary theoretical justification for the observed high correlation between TM-score and I-INF.

Attachment

Submitted filename: response-to-referee.pdf

pone.0319917.s004.pdf (34.8KB, pdf)

Decision Letter 2

Yong Wang

11 Feb 2025

Assessing interface accuracy in macromolecular complexes

PONE-D-24-53421R2

Dear Dr. Szachniuk,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Yong Wang

Academic Editor

PLOS ONE

Acceptance letter

Yong Wang

PONE-D-24-53421R2

PLOS ONE

Dear Dr. Szachniuk,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Yong Wang

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Table. Evaluation of the predicted 3D models from the RNA-protein docking decoys.

    (PDF)

    pone.0319917.s001.pdf (265.2KB, pdf)
    S1 Fig. A distribution of TM-score, DockQv2, and I-INF values computed for the benchmark set.

    (PDF)

    pone.0319917.s002.pdf (91.7KB, pdf)
    Attachment

    Submitted filename: Response-to-reviewers.pdf

    pone.0319917.s003.pdf (89.1KB, pdf)
    Attachment

    Submitted filename: response-to-referee.pdf

    pone.0319917.s004.pdf (34.8KB, pdf)

    Data Availability Statement

    The tool for computing I-INF and F1 to assess interfaces in macromolecular assemblies, along with a user manual and ready-to-run examples, is publicly available at Zenodo (https://dx.doi.org/10.5281/ zenodo.14697284) and GitHub (https://github.com/OlgierdL/iinf). For benchmarking, we used the dataset available at https://zoulab.dalton.missouri.edu/RNAdecoys/index.html.


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES