Skip to main content
Sensors (Basel, Switzerland) logoLink to Sensors (Basel, Switzerland)
. 2026 Jan 20;26(2):688. doi: 10.3390/s26020688

NTFold: Structure-Sensing Nucleotide Attention Learning for RNA Secondary Structure Prediction

Kangjun Jin 1, Zhuo Zhang 1,*, Guipeng Lan 1, Shuai Xiao 1, Jiachen Yang 1
Editors: Tibor Hianik1, Almira Ramanaviciene1
PMCID: PMC12845834  PMID: 41600484

Abstract

Determining RNA secondary structures is a fundamental challenge in computational biology and molecular sensing. Experimental techniques such as X-ray crystallography, nuclear magnetic resonance, and cryo-electron microscopy can reveal RNA structures with atomic precision, but their high cost and time consuming nature limit large-scale applications. To address this issue, we introduce the Structure-Sensing Nucleotide Attention Learning framework (NTFold), a virtual sensing framework based on deep learning for accurate RNA secondary structure prediction. NTFold integrates a Nucleotide Attention Module (NAM) to explicitly model dependencies among nucleotides, thereby capturing fine-grained sequence correlations. The resulting correlation map is subsequently refined by a Structural Refinement Module (SRM), which preserves hierarchical spatial information and enforces structural consistency. Through this two stage learning paradigm, NTFold produces high-precision contact maps that enable reliable RNA secondary structure reconstruction. Extensive experiments demonstrate that NTFold outperforms existing deep learning-based predictors, highlighting its capability to learn both local and global nucleotide interactions in an sensor inspired manner. This study provides a new direction for integrating attention driven correlation modeling with structure-sensing refinement toward efficient and scalable RNA structural sensing.

Keywords: RNA secondary structure prediction, nucleotide attention mechanism, structure-sensing refinement

1. Introduction

RNA secondary structure prediction is a fundamental challenge in computational biology and is central to understanding RNA function, regulatory mechanisms, and cellular processes [1,2]. As a versatile biological macromolecule, RNA participates in a wide range of cellular activities, including gene expression regulation, catalysis, ligand sensing, and chromatin organization. Beyond its roles in modern cellular systems, RNA is also thought to have played a pioneering role in early life, where its structural versatility enabled both information storage and functional activity, supporting evolutionary plausibility under primitive conditions [3]. Unlike proteins, whose functionality is primarily encoded in the three-dimensional structure, RNA function is deeply intertwined with its secondary structure, which defines the canonical and non-canonical base-pairing patterns, as illustrated in Figure 1. These paired and unpaired regions serve as the architectural foundation for RNA tertiary folding, molecular recognition, and interactions with proteins, metabolites, and other nucleic acids [4]. Therefore, accurate secondary structure prediction is indispensable for interpreting RNA biology and for enabling RNA-design applications in synthetic biology and therapeutics.

Figure 1.

Figure 1

The graph and matrix representation of an RNA secondary structure.

In addition to general RNA secondary structure prediction, the modeling of pseudoknotted structures has long been recognized as a distinct and challenging problem. Early pseudoknot-oriented methods, such as HotKnots [5], extend thermodynamic folding frameworks by heuristically exploring low-energy structures that include crossing base pairs, while IPknot [6] formulates pseudoknot prediction as an integer programming problem to identify globally consistent base-pairing configurations. These approaches highlight the algorithmic complexity and diversity of strategies required to address different classes of pseudoknots. In this context, learning-based frameworks such as NTFold aim to model pseudoknotted interactions at the contact map level, providing a complementary perspective that emphasizes interaction pattern learning rather than explicit combinatorial optimization.

Despite the importance of RNA structure, determining RNA conformations experimentally remains a costly and low-throughput endeavor. Experimental techniques such as nuclear magnetic resonance (NMR) spectroscopy [7] and cryo-electron microscopy (cryo-EM) [8] can provide high-resolution structures, but they require sophisticated instrumentation, significant time investment, and often struggle with conformational heterogeneity or long and flexible RNAs. Consequently, the number of experimentally resolved RNA structures in the Protein Data Bank is orders of magnitude smaller than that of proteins, and is heavily biased toward small, stable, or highly conserved RNAs. This scarcity hinders the comprehensive characterization of structural diversity across RNA families, further emphasizing the need for robust computational prediction methods.

Computational approaches for RNA secondary structure prediction have traditionally relied on dynamic programming (DP) algorithms grounded in thermodynamic modeling. Classical energy-based methods assume that the native RNA structure corresponds to the minimum free energy (MFE) conformation and employ nearest-neighbor thermodynamic parameters to evaluate base-pairing and loop energetics [9,10]. Representative algorithms include the Zuker algorithm [11], as implemented in RNAfold [12], as well as Mfold [13] and the RNAstructure suite [14], which enumerate large numbers of candidate secondary structures and identify the MFE solution while also estimating base-pairing probabilities. Although these models have been pervasive and have served as the gold standard for decades, they suffer from several inherent limitations. First, their performance depends strongly on the completeness and accuracy of experimentally derived thermodynamic parameter sets, which remain incomplete for many complex motifs and non-canonical interactions. Second, the cubic time complexity of DP-based algorithms makes them difficult to scale to long RNA sequences and large conformational spaces. Third, structurally distinct RNA folds may exhibit near-identical free-energy values, rendering the MFE criterion ambiguous in practice. Moreover, most DP-based algorithms cannot efficiently handle pseudoknots due to prohibitive time and space complexity, leading to their routine exclusion despite biological relevance [15]. These limitations motivate alternative, data-driven approaches that infer RNA secondary structures directly from sequence patterns without explicit free-energy minimization. In contrast, NTFold adopts a data-driven learning paradigm that does not rely on predefined thermodynamic parameters or explicit energy minimization. Instead of searching for a single optimal structure, NTFold infers base-pairing patterns directly from sequence-derived interaction representations. The proposed Structural Refinement Module enforces structural consistency by reinforcing mutually exclusive base-pairing patterns and coherent helices, implicitly respecting Watson–Crick pairing rules while allowing complex interaction topologies, including pseudoknots, to be naturally represented in the predicted contact map. This paradigm shift from energy-based enumeration to pattern-driven inference highlights the novelty of NTFold and its potential advantages for scalable RNA secondary structure prediction.

Comparative sequence analysis methods provide an alternative approach by leveraging co-evolutionary information from aligned homologous sequences to identify conserved structural elements. These methods, typically formulated using stochastic context-free grammars (SCFGs) [16], exploit compensatory mutations as strong evidence for base-pairing. More recently, pipelines such as RNAcmap construct RNA contact maps by extracting co-evolutionary signals from multiple sequence alignments, providing an evolutionary constraint-driven alternative to single-sequence structure prediction [17]. While powerful in principle, their performance is fundamentally limited by the availability of sufficiently large and diverse homologous sequence families. Given that the number of well-characterized RNA families is small and unevenly distributed across organisms, comparative methods often fail when applied to novel or poorly annotated RNA sequences.

The rapid development of deep learning (DL) has transformed numerous fields, including computer vision [18,19,20], image assessment [21,22], remote sensing [23], and Internet of things [24,25]. Motivated by these successes, the application of deep learning to RNA secondary structure prediction has gained considerable momentum. Early models such as SPOT-RNA [26] integrate convolutional neural networks and bidirectional LSTMs to learn sequence-to-structure mappings. CDPFold [27] merges DL feature extraction with classical DP decoding to enhance prediction stability. Ufold [28] employs a U-Net architecture to predict contact maps by treating RNA structure prediction as an image segmentation task. E2Efold [29] and RFold [30] embed hard structural constraints directly into their unrolled optimization frameworks, thereby enforcing structural validity during decoding. MXFold2 [31] adopts multi-scale convolutional representations to capture hierarchical base-pairing patterns, while DSRNAFold [32] enhances prediction by integrating deep learning with structural context analysis. Despite these advances, many existing DL-based methods primarily emphasize feature extraction or constrained decoding, with structural validity often enforced either implicitly or at specific stages of the pipeline. In contrast, NTFold is designed to explicitly bridge data-driven interaction learning and structure-aware refinement. Rather than treating attention mechanisms solely as feature encoders, NTFold leverages attention-derived correlation maps as an intermediate structural representation and applies a dedicated Structural Refinement Module to enforce global consistency, mutual exclusivity of base-pairing, and coherent helix formation. This explicit coupling of interaction learning and structural refinement distinguishes NTFold from prior deep learning approaches and provides a more interpretable and scalable framework for RNA secondary structure prediction.

Among recent methodological advances, attention mechanisms have become a powerful paradigm for modeling interaction dependencies and have achieved remarkable success in computer vision [33], medical image analysis [34,35], and biomedical applications [34,35]. Motivated by their strong capability in capturing long-range relational patterns, several RNA secondary structure prediction models have incorporated attention mechanisms to enhance the modeling of base-pairing interactions. For example, ATTfold explicitly integrates attention layers with structural constraints to model long-range dependencies, including pseudoknotted interactions [36], while other approaches embed attention within end-to-end prediction pipelines [29]. However, most existing attention-based approaches treat nucleotide interactions implicitly within transformer-style architectures, where attention primarily serves as a feature reweighting mechanism rather than an explicit structural representation. As a result, nucleotide–nucleotide dependencies are not directly formulated or refined as interpretable interaction maps, which may obscure fine-grained relational cues embedded in primary sequences and limit the extraction of biologically meaningful correlation patterns. In contrast, NTFold elevates attention-derived correlation maps to a central structural representation, which are subsequently refined to enforce global consistency and mutually exclusive base-pairing patterns, thereby providing a clear methodological departure from prior attention-based RNA secondary structure predictors.

In contrast, this work takes a more biologically grounded perspective by designing the model as a functional analogue of a structural sensor that directly perceives and interprets nucleotide relationships. To this end, a Nucleotide Attention Module (NAM) is introduced to explicitly construct a high-resolution interaction map by modeling how each nucleotide responds to and influences others within the sequence. This produces an interpretable, attention-driven correlation representation that reflects the underlying biochemical dependency patterns. Building upon this representation, we further develop a Structural Refinement Module (SRM) to refine the initial interaction map into a biologically coherent contact matrix, leveraging hierarchical structural features. Together, these components allow the model to function as a biologically inspired sensing system that detects, integrates, and interprets nucleotide interactions, ultimately enabling accurate RNA secondary structure prediction. Our contributions are as follows:

(1) We introduce NTFold, a virtual sensing framework that enables efficient and scalable RNA secondary structure prediction, providing a computational alternative to costly and low-throughput experimental structure determination.

(2) We design a Nucleotide Attention Module that explicitly captures nucleotide dependencies, enabling fine-grained sensing of sequence-level correlations.

(3) We develop a structure refinement module that enhances spatial consistency and improves the accuracy of the predicted contact map, enabling more reliable RNA secondary structure reconstruction.

2. Related Work

2.1. Thermodynamic Models

Traditional methods for RNA secondary structure prediction are primarily based on thermodynamic principles, with the core idea of finding the structure with the Minimum Free Energy (MFE). The pioneering work of Zuker and Nussinov laid the foundation for this field, proposing dynamic programming algorithms that can find the optimal non-pseudoknotted structure in polynomial time [11,37]. Mfold [13] is a widely used web-based RNA and DNA secondary structure prediction tool that employs thermodynamic free-energy minimization to generate candidate folding structures and related stability analyses from single-stranded nucleic acid sequences. Tools such as the ViennaRNA package (RNAfold) [12] and RNAstructure [14] are widely used, leveraging experimentally determined thermodynamic parameters to evaluate and predict structural stability. However, these methods face two major challenges: first, their computational complexity (typically O(N3)) becomes prohibitive for long RNA sequences; second, they struggle to handle complex pseudoknotted structures, with complexity increasing to O(N6) or higher, limiting their application in genome-scale analysis.

2.2. Hybrid Methods

Hybrid methods combine the advantages of multiple strategies to overcome the limitations of single methods. RNAalifold, an early hybrid method in the ViennaRNA package, combines energy-based and comparative methods. It calculates the minimum energy structure of a set of aligned input sequences by modifying the scoring scheme of the dynamic programming algorithm in traditional thermodynamic methods [38]. CentroidFold is based on the γ-centroid estimator and supports multiple probabilistic models, providing more accurate predictions [39]. MXFold integrates thermodynamic information with structured support vector machines (SSVMs) and predicts the optimal secondary structure through a Zuker-style dynamic programming algorithm, improving the prediction accuracy and robustness [40]. MXFold2 further improves on this basis. It replaces SSVMs with deep neural networks, combines CNN and BiLSTM layers to learn folding parameters, and calculates the final score through a dynamic programming algorithm, showing excellent performance on multiple test datasets [31].

2.3. Deep Learning-Based Methods

Deep learning-based approaches have gained significant traction in RNA secondary structure prediction due to their ability to model complex relationships between RNA sequences and their structural features. SPOT-RNA [26] combines ResNet and bidirectional LSTM architectures, leveraging transfer learning to predict all base pairs, including non-canonical and pseudoknotted ones. CDPFold [27] integrates convolutional neural networks (CNNs) with dynamic programming to enhance prediction accuracy. E2Efold [29] is an end-to-end deep learning model that has made significant progress in pseudoknot prediction. MXFold2 [31] combines deep learning with thermodynamic models, employing a multi-scale convolutional neural network to handle long sequences and capture complex base-pairing patterns [31]. Ufold [28] employs a U-Net architecture to represent RNA sequences as “images” for structural inference. RFold [30] proposes a framework that leverages solutions to the K-Rooks problem to guide RNA secondary structure prediction, ensuring structural constraints are met during inference. DSRNAFold [32] combines deep learning techniques with structural context analysis to improve RNA secondary structure prediction. These methods collectively demonstrate the potential of deep learning in advancing RNA structural modeling.

3. Method

3.1. Overview of the Proposed Framework

In this study, we propose an efficient deep learning framework for RNA secondary structure prediction, which directly maps nucleotide sequences to base-pair contact maps. The overall pipeline consists of three key components: (1) A NAM for constructing pairwise correlation maps; (2) A SRM for enhancing structural consistency; (3) A post-processing module for generating the final contact map. The complete workflow is illustrated in Figure 2.

Figure 2.

Figure 2

The overall framework of NTFold.

Given an input RNA sequence, each nucleotide is first encoded as a learnable embedding vector to capture local sequence context. The embedding sequence is then processed by a multi-head self-attention module to model global dependencies and infer intrinsic pairwise interaction tendencies. Aggregating the attention weights yields an initial correlation map, which serves as an initial interaction map.

Despite capturing global interactions, the raw attention map often contains noise and weak long-range signals. To address these issues, we employ a structural refinement module, which extracts multi-scale features and reconstructs the interaction map. This process strengthens canonical and non-canonical pairing patterns and suppresses spurious correlations, resulting in a refined interaction map that better reflects RNA structural principles.

Finally, the contact map generation module applies structural constraints to produce a binary base-pair contact map. Overall, our framework integrates global dependency modeling via attention with spatial refinement through structural refinement module, enabling accurate and interpretable RNA secondary structure prediction without reliance on manual features or thermodynamic models.

3.2. Nucleotide Embedding

Given an RNA sequence x={x1,x2,,xL} of length L, we first convert each nucleotide into a learnable embedding vector. This step transforms discrete categorical symbols into a continuous representation space that captures their biochemical properties and contextual significance. Formally, we define a trainable embedding matrix ER|V|×d, where V={A,U,G,C,N} denotes the nucleotide embedding and d is the feature dimension. The initial embedding is computed by

ei=E(xi),i=1,2,,L. (1)

To incorporate positional dependency, which is essential for capturing non-local interactions and structural motifs, we add a deterministic positional encoding term pi to each embedding:

zi=ei+pi, (2)

where pi encodes the relative order along the RNA chain, enabling the model to distinguish nucleotides with identical identities but different positions. The resulting embedding matrix is

Z=[z1;z2;;zL]RL×d. (3)

This embedding representation serves as the input feature backbone for downstream attention inference and structural refinement. It ensures that both nucleotide identity and positional context are preserved, providing a biologically meaningful representation for subsequent interaction modeling.

3.3. Nucleotide Attention Module

To explicitly model long-range nucleotide dependencies and infer potential pairing tendencies, we employ a multi-head self-attention mechanism that transforms the embedding features into an interaction-aware representation. Given the embedding matrix

ZRL×d,

we compute, for each attention head h, a query, key, and value triplet through learned linear projections:

Qh=ZWhQ,Kh=ZWhK,Vh=ZWhV, (4)

where WhQ,WhK,WhVRd×dh are trainable matrices and dh denotes the representation dimension per head.

The interaction strengths between nucleotide positions i and j are quantified through scaled dot-product attention:

Ah=SoftmaxQhKhdh, (5)

where Ah(i,j) reflects how strongly nucleotide xi attends to nucleotide xj. To obtain a unified correlation estimate, we average all attention heads:

C=1Hh=1HAh,CRL×L. (6)

The resulting matrix C provides a coarse estimation of intrinsic nucleotide interaction likelihoods, capturing canonical pairing patterns (e.g., A–U, C–G) as well as non-canonical structural relationships over long distances.

Unlike sequential convolution, attention enables direct global receptive fields without increasing model depth, allowing distal nucleotides involved in stems, pseudoknots, or long hairpins to influence each other immediately. Thus, C forms the initial structural correlation map that bridges sequence encoding with spatial structural inference, serving as the input to the refinement module for spatially coherent reconstruction.

3.4. Structural Refinement Module

The correlation map CRL×L obtained from the Nucleotide Attention Module provides a coarse estimate of base-pairing propensities. However, it often contains noise, fragmented helices, and weak long-range signals. To address these issues, we introduce a Structural Refinement Module, which refines C into a more spatially coherent interaction map Crefined.

The refinement module is implemented as a U-Net style encoder–decoder network. The encoder progressively extracts multi-scale features:

Fk=ConvBlockkDown(Fk1),F0=C, (7)

where k=1,,n, Down(·) denotes a downsampling operation, and Fk captures structural features at different nucleotide scales, with shallow layers focusing on local motifs and loop-level structures within approximately 30 nucleotides, and deeper layers modeling global structural context characterized by long-range dependencies spanning hundreds of nucleotides.

The decoder restores spatial resolution while integrating skip connections from the encoder:

Gk=ConvBlockkUp(Gk1)Fnk, (8)

where Up(·) denotes upsampling, ‖ represents feature concatenation, and n is the total number of encoder/decoder layers. Canonical pairing patterns are strengthened due to their consistently high attention scores and spatial continuity in the correlation map, whereas non-canonical pairs are retained when they form stable, context-supported interaction patterns across adjacent positions. The refinement process favors such recurrent structural signals while filtering out noise-like, unsupported interactions.

Finally, the refined correlation map is obtained as

Crefined=SRM(C),CrefinedRL×L. (9)

The Structural Refinement Module thus produces a high-quality interaction map that better respects RNA structural principles, serving as the input to the post-processing module for binary contact map generation. By integrating multi-scale local and global features, this module strengthens helix continuity, improves structural consistency, and enhances the reliability of downstream RNA secondary structure reconstruction.

3.5. Post-Processing Module

The refined correlation map CrefinedRL×L produced by the Structural Refinement Module encodes base-pairing propensities with reduced noise and improved helix continuity. To convert this map into a biologically valid RNA secondary structure, we apply a post-processing step that enforces structural constraints and generates the final binary contact map M{0,1}L×L.

We first impose a sparsity-aware thresholding on Crefined using a smooth approximation of the sign function:

C˜=soft_sign(Crefineds)Crefined, (10)

where s is a predefined threshold and ⊙ denotes element-wise multiplication. This step suppresses weak interactions while retaining strong base-pair signals.

Next, we enforce Lagrangian constraints to satisfy RNA pairing rules. Let A denote the contact probabilities after constraint optimization. The update rule for A is

G=Aλ·soft_sign(jAij1)12A, (11)
AAηG, (12)
λλ+ηλ·ReLUjAij1, (13)

where λ is a Lagrange multiplier and η and ηλ are learning rates for the minimization and maximization steps, respectively, and the sum is taken over each nucleotide to ensure that each pairs with at most one partner.

Finally, we generate the binary contact map M by applying the following structural constraints: (1) Symmetry: M=M; (2) Pairing exclusivity: Each nucleotide pairs with at most one partner; (3) Local masks: Prohibit impossible pairings due to minimum loop lengths or steric constraints.

The resulting map M represents a valid RNA secondary structure, ready for downstream analysis or visualization. This post-processing module integrates constraint optimization with biologically motivated rules, ensuring that the final predictions are both accurate and structurally plausible.

4. Experiments

4.1. Dataset Selection and Analysis

To comprehensively evaluate the proposed model, we adopt three widely used RNA secondary structure datasets as benchmarks, including RNAStrAlign [41], bpRNA-1m [42], and ArchiveII [43]. These datasets cover diverse RNA families and a broad range of sequence lengths, providing a solid foundation for assessing model accuracy, robustness, and generalization performance.

RNAStrAlign: RNAStrAlign [41] contains 30,451 RNA secondary structures spanning eight RNA families. After applying quality control and sequence-length filtering, we obtain a refined training set containing 20,862 high-quality RNA samples. The distribution of sequence lengths and sample counts across families is summarized in Table 1. This analysis also reveals distinct sparsity characteristics among different RNA families, which is crucial for evaluating the model’s capability in handling variable sparsity patterns.

Table 1.

Analysis on RNAStralign training set.

Type Length Samples
16SrRNA 54∼600 1226
5SrRNA 104∼132 9133
tRNA 59∼95 7401
grp1 163∼600 1683
SRP 30∼553 493
tmRNA 102∼437 514
RNaseP 189∼486 379
telomerase 382∼559 33

ArchiveII: ArchiveII [43] contains 3975 experimentally validated RNA secondary structures and has been widely adopted as a benchmark for RNA secondary structure prediction. In this work, we use ArchiveII exclusively for evaluating the model trained on RNAStrAlign, thereby assessing its generalization ability across datasets.

bpRNA-1m: For the bpRNA-1m dataset [42], which contains 102,318 sequences from 2588 families, we followed the pre-processing protocol of MXfold2 [31] by using CD-HIT to remove redundant sequences, and then randomly divided the remaining data into training and testing subsets, denoted as TR0 and TS0.

4.2. Evaluation

We evaluate model performance using precision, recall, and F1 score. Precision measures the proportion of correctly predicted base pairs among all predicted base pairs, while recall quantifies the proportion of true base pairs that are successfully recovered. The F1 score is defined as the harmonic mean of precision and recall, providing a balanced evaluation of base-pair prediction accuracy that accounts for both false positives and false negatives. These metrics are computed as follows:

Precision=TPTP+FP, (14)
Recall=TPTP+FN, (15)
F1=2×Precision×RecallPrecision+Recall, (16)

where TP denotes the number of correctly predicted base pairs, FP represents the number of falsely predicted base pairs, and FN indicates the number of missed true base pairs.

4.3. Standard RNA Secondary Structure Prediction

Table 2 presents the performance comparison of different RNA secondary structure prediction methods on the RNAStrAlign test set. Traditional thermodynamic approaches such as Mfold, RNAfold, and RNAstructure achieve relatively low performance, with F1 scores ranging from 0.42 to 0.55. With the incorporation of statistical learning and efficient inference strategies, methods like CONTRAfold, LinearFold, and CDPfold show moderate improvements, reaching F1 values around 0.61–0.63. Deep learning-based approaches demonstrate a substantial performance gain, where UFold achieves an F1 score of 0.915, indicating stronger capability in detecting true base-pairing patterns. In comparison, our proposed model achieves the best overall performance, obtaining 0.979 in Precision, 0.976 in Recall, and 0.977 in F1 score. This represents an improvement of approximately 6.2% over UFold in terms of F1, highlighting the superiority of our method in accurately capturing base-pair dependencies and delivering more reliable structural predictions.

Table 2.

Results on RNAStralign test set. The best results are highlighted in bold, and the second-best results are underlined.

Method Precision Recall F1
Mfold [13] 0.450 0.398 0.420
RNAfold [12] 0.516 0.568 0.540
RNAstructure [44] 0.537 0.568 0.550
CONTRAfold [45] 0.608 0.663 0.633
LinearFold [46] 0.620 0.606 0.609
CDPfold [27] 0.633 0.597 0.614
E2Efold [29] 0.866 0.788 0.821
UFold [28] 0.905 0.927 0.915
Ours 0.979 0.976 0.977

4.4. Evaluating Model Generalization

Table 3 reports the performance of various RNA secondary structure prediction approaches on the ArchiveII dataset. Traditional thermodynamic models such as Mfold, RNAfold, RNAstructure, and RNAsoft achieve moderate performance, with F1 scores generally ranging from 0.60 to 0.65. Methods incorporating probabilistic modeling or structural constraints, including CONTRAfold, LinearFold, and Contextfold, yield notable gains, with Contextfold reaching an F1 score of 0.842, demonstrating the advantage of more expressive modeling strategies. Deep learning-based approaches exhibit substantial improvements over classical methods. Early neural frameworks, such as E2Efold and SPOT-RNA, already show clear performance advantages, while more recent architectures including MXfold2, RTfold, UFold, DSRNAFold, and Sincfold achieve further improvements in both precision and recall. Notably, Sincfold attains the highest recall 0.953 and an F1 score of 0.923, indicating strong capability in capturing true base-pairing interactions. Our proposed method achieves the best overall performance, obtaining the highest precision of 0.941 and the highest F1 score of 0.931, while maintaining competitive recall of 0.933. This demonstrates that our model yields more accurate base-pair predictions with fewer false positives, while preserving strong sensitivity to true interactions.

Recent studies have shown that the generalization ability of deep learning-based RNA secondary structure predictors is strongly influenced by sequence similarity and RNA family distribution between training and test sets. In particular, Qiu et al. demonstrated that apparent performance gains may diminish under distribution shifts when homologous or structurally related sequences are excluded, highlighting sequence similarity as a key determinant of generalizability [47]. In this context, the dataset splits used in our evaluation are designed to mitigate information leakage across RNA families, providing a more stringent assessment of out-of-distribution performance. More broadly, as summarized in the review by Zhao et al., machine learning-based RNA secondary structure predictors can be categorized by their modeling assumptions, input representations, and decoding strategies, each exhibiting distinct generalization behaviors across RNA families [48]. Our results should therefore be interpreted within this framework, where improved performance reflects both architectural design and the ability to capture transferable structural patterns beyond sequence-level similarity. Overall, these results confirm that our approach establishes a new state-of-the-art on the ArchiveII benchmark, outperforming classical, probabilistic, and deep learning-based predictors, thereby validating the effectiveness and strong generalization capability of our proposed framework.

Table 3.

Results on ArchiveII dataset. The best results are highlighted in bold, and the second-best results are underlined.

Method Precision Recall F1
Mfold [13] 0.668 0.590 0.621
CDPfold [27] 0.557 0.535 0.545
RNAfold [12] 0.663 0.613 0.631
RNAstructure [44] 0.664 0.606 0.628
CONTRAfold [45] 0.696 0.651 0.665
LinearFold [46] 0.724 0.605 0.647
RNAsoft [49] 0.665 0.694 0.622
E2Efold [29] 0.734 0.660 0.686
SPOT-RNA [26] 0.743 0.726 0.711
MXfold2 [31] 0.788 0.760 0.768
Contextfold [50] 0.873 0.821 0.842
RTfold [51] 0.891 0.789 0.814
UFold [28] 0.887 0.928 0.905
DSRNAFold [32] 0.896 0.927 0.907
Sincfold [52] 0.908 0.953 0.923
Ours 0.941 0.933 0.931

4.5. Large-Scale Benchmark Evaluation

Table 4 reports the performance of mainstream RNA secondary structure prediction approaches on the bpRNA-TS0 test set, which is constructed under the TR0–TS0 separation strategy to evaluate true generalization to unseen structural families. Traditional thermodynamic models, such as RNAfold, Mfold, and RNAstructure, yield moderate performance with F1 scores around 0.53–0.54. Methods incorporating machine learning priors, such as CONTRAfold and Contextfold, achieve slightly higher accuracy, though improvements remain limited. Among early neural models, SPOT-RNA achieves the highest recall of 0.693, indicating strong sensitivity to true base-pairing signals, yet its precision remains moderate, reflecting a higher false positive rate. Recent data-driven approaches demonstrate further enhancements. DSRNAFold shows a notable improvement in precision of 0.641 and achieves an F1 score of 0.627, reflecting better structural consistency. Sincfold obtains the highest precision of 0.702, but its lower recall reduces its ability to fully recover true interactions, limiting overall performance. Our method achieves the highest F1 score of 0.645, establishing the best balance between precision and recall among all evaluated models. Compared to MXfold2—the baseline corresponding to this dataset split—our method improves precision and recall by 11.3% and 6.7%, respectively. This indicates that our approach not only reduces structural false positives but also improves the recovery of true base pairs. Overall, the results on bpRNA-TS0 confirm that our method demonstrates superior robustness and generalization under distribution shifts, effectively handling previously unseen RNA structural patterns.

Table 4.

Results on bpRNA-TS0 set. The best results are highlighted in bold, and the second-best results are underlined.

Method Precision Recall F1
E2Efold [29] 0.140 0.129 0.130
RNAstructure [44] 0.494 0.622 0.533
RNAsoft [49] 0.497 0.626 0.535
RNAfold [12] 0.494 0.631 0.536
Mfold [13] 0.501 0.627 0.538
Contextfold [50] 0.529 0.607 0.546
LinearFold [46] 0.561 0.581 0.550
MXfold2 [31] 0.519 0.646 0.558
CONTRAfold [45] 0.528 0.655 0.567
SPOT-RNA [26] 0.594 0.693 0.619
UFold [28] 0.521 0.588 0.553
DSRNAFold [32] 0.641 0.649 0.627
Sincfold [52] 0.702 0.604 0.624
Ours 0.632 0.689 0.645

4.6. Pseudoknot Prediction

Table 5 presents pseudoknot prediction performance on the RNAStrAlign benchmark. Classical thermodynamic approaches such as RNAstructure achieve reasonable overall accuracy F1 of 0.769, yet their recall is limited, indicating challenges in capturing complex base-pairing patterns. Reinforced by deep learning, SPOT-RNA substantially improves recall by 0.978, but at the expense of precision, reflecting over-prediction tendencies. More recent end-to-end neural architectures, including E2Efold and UFold, further boost both metrics, with UFold achieving strong balance F1 of 0.976. Our method achieves the best results overall, obtaining the highest precision of 0.980 and the highest F1 score of 0.977, demonstrating superior capability in accurately identifying pseudoknots while minimizing false positives. These results highlight that our approach offers more reliable structural modeling, particularly for complex RNA folding patterns that are known to be challenging for existing predictors.

Table 5.

Pseudoknot prediction results on RNAStralign dataset. The best results are highlighted in bold, and the second-best results are underlined.

Method Precision Recall F1
RNAstructure [44] 0.778 0.761 0.769
SPOT-RNA [26] 0.677 0.978 0.800
E2Efold [29] 0.844 0.990 0.911
UFold [28] 0.962 0.990 0.976
Ours 0.980 0.976 0.977

4.7. Ablation Experiment

The results in Table 6 clearly highlight the effectiveness of both the Nucleotide Attention Module and the Refinement Block within our model architecture. When either component is removed, the performance consistently drops across all three datasets, demonstrating their complementary contributions.

Table 6.

Ablation experiment results of RNAStralign, ArchiveII, and bpRNA-TS0 datasets. The best results are highlighted in bold. The w/o denotes the variant without the corresponding module.

RNAStralign Dataset
Method Precision Recall F1
NTFold 0.979 0.976 0.977
w/o Nucleotide Attention 0.926 0.938 0.927
w/o Refinement Block 0.951 0.961 0.953
ArchiveII Dataset
Method Precision Recall F1
NTFold 0.941 0.933 0.931
w/o Nucleotide Attention 0.912 0.906 0.911
w/o Refinement Block 0.923 0.918 0.919
bpRNA-TS0 Dataset
Method Precision Recall F1
NTFold 0.632 0.689 0.645
w/o Nucleotide Attention 0.601 0.657 0.623
w/o Refinement Block 0.622 0.667 0.631

First, removing the Nucleotide Attention Module causes the most noticeable degradation. For example, on the RNAStralign dataset, the F1 score decreases from 0.977 to 0.927, showing that attention significantly enhances base-pair interaction modeling by explicitly capturing nucleotide-wise contextual dependencies. A similar trend is observed on ArchiveII and bpRNA-TS0, where F1 drops from 0.931 to 0.911 and from 0.645 to 0.623, respectively. This indicates that nucleotide-level attention is crucial for learning more accurate long-range pairing relations and improving prediction robustness across diverse RNA families.

Second, removing the Refinement Block also results in performance decline, though to a lesser extent compared with removing attention. For RNAStralign, the F1 decreases from 0.977 to 0.953, and similar performance gaps appear on ArchiveII and bpRNA-TS0. These results confirm that the Refinement Block further improves the initially predicted interaction map by suppressing noise and enhancing structural consistency, leading to more precise base-pair confidence estimation.

Overall, the full model consistently achieves the best results under all evaluation metrics. The complementary effects of Nucleotide Attention in capturing fine-grained dependencies and the Refinement Block in boosting structural refinement collectively drive the strong performance of our approach.

4.8. Inference Time Comparison

The inference time comparison in Table 7 demonstrates that our proposed method achieves highly efficient RNA structure prediction while maintaining strong accuracy. For short sequences of 128 nucleotides, our model completes inference in only 0.032 s per sample, making it the fastest among all existing GPU-based predictors. Compared with widely used deep learning methods such as MXfold2 0.204s, SPOT-RNA 0.847s, and SincFold 0.267s, our approach reduces latency by a significant margin, indicating a more streamlined computation pipeline. When handling longer RNA sequences 512 nt, our method retains the same efficiency advantage. The inference time increases only slightly to 0.154 s per sequence, which is still substantially lower than MXfold2 3.117s, SPOT-RNA 1.877s, and DSRNAFold 0.756s. Even relative to UFold, a known efficient model, our method achieves approximately twice the speed improvement. This stable scalability across sequence lengths suggests that our design avoids excessive computation growth when modeling long-range dependencies. It is worth noting that, beyond learning-based predictors, computational efficiency in RNA folding has also been addressed through algorithmic approximations within the thermodynamic framework. Representative methods such as LinearPartition approximate ensemble partition functions in linear time, offering an alternative notion of tractability by trading exactness for scalability [53]. While such approaches do not directly compete with single-sequence structure predictors or contact map-based models, they provide an important complementary baseline for understanding speed–accuracy trade-offs in RNA structure analysis. In summary, the results show that our model simultaneously achieves high prediction accuracy and fast inference speed. The reduced per sequence runtime and favorable scalability make it suitable for large-scale RNA datasets and real-time downstream biological applications.

Table 7.

Inference time on RNAStralign test set.

Method Time/Seq (=128 nt) Time/Seq (=512 nt) Device
CycleFold [54] 4.117 s 57.927 s CPU
ProbKnot [15] 0.261 s 4.332 s CPU
LinearPartition [53] 0.103 s 0.602 s CPU
CONTRAfold [45] 0.043 s 0.531 s CPU
LinearFold [46] 0.087 s 0.306 s CPU
RNAfold [12] 0.024 s 0.166 s CPU
MXfold2 [31] 0.204 s 3.117 s GPU
SPOT-RNA [26] 0.847 s 1.877 s GPU
SincFold [52] 0.267 s 2.661 s GPU
DSRNAFold [32] 0.056 s 0.756 s GPU
UFold [28] 0.041 s 0.346 s GPU
Ours 0.032 s 0.154 s GPU

5. Discussion

The experimental results indicate that NTFold not only improves predictive accuracy but also recovers biologically meaningful RNA secondary structure patterns. Analysis of the predicted contact maps shows that the model consistently captures canonical base-pairing signals and forms continuous stem regions, while loop boundaries are clearly distinguished by attenuated interaction strengths, reflecting known RNA folding rules. Structural comparisons across RNAs from the same families further suggest that conserved stem–loop organizations are preserved even under sequence variation, indicating that the model learns family-level structural regularities rather than relying solely on local sequence features. Compared with previous deep learning models such as convolutional or U-Net-based predictors, the explicit nucleotide attention mechanism in NTFold provides a clearer representation of long-range dependencies and yields interpretable correlation maps that reveal nucleotide interaction tendencies. The ablation studies support this interpretation: removing the attention module weakens long-range base-pair recovery, while excluding the refinement module leads to fragmented helices and increased spurious contacts, confirming their complementary roles in capturing global dependencies and enforcing structural coherence. Although challenges remain in modeling complex tertiary interactions and multiple metastable conformations, these results demonstrate that NTFold aligns well with established RNA structural principles and offers both improved performance and mechanistic interpretability.

6. Conclusions

In this work, we present a data-driven framework for RNA secondary structure prediction designed for molecular sensing and diagnostic applications. The model combines nucleotide-level attention, spatial refinement, and constraint-driven decoding to generate accurate and biologically consistent RNA contact maps directly from sequences, without relying on thermodynamic folding algorithms. By capturing global interaction tendencies through attention-derived correlation maps and enhancing structural continuity via refinement, the framework improves the detectability of motifs such as stems and loops that are critical for RNA-based sensors. Constraint-driven decoding ensures biologically plausible structures, making the predictions applicable to molecular engineering tasks. The method demonstrates competitive accuracy, interpretability, computational efficiency, and modularity, supporting integration into pipelines for aptamer design, CRISPR guide optimization, and analysis of target-induced structural changes. Overall, this framework bridges deep learning-based structural modeling with practical RNA sensing needs, offering both methodological innovations and applied value for biosensing research.

Acknowledgments

We would like to thank all those who have contributed to the work of this paper.

Author Contributions

Conceptualization, K.J., Z.Z. and J.Y.; methodology, K.J. and G.L.; software, K.J. and S.X.; validation, G.L. and Z.Z.; formal analysis, S.X.; investigation, Z.Z.; resources, S.X.; data curation, K.J.; writing—original draft preparation, K.J.; writing—review and editing, G.L.; visualization, Z.Z.; supervision, J.Y.; project administration, S.X.; funding acquisition, S.X. and J.Y. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data were derived from the following resources available in the public domain: RNAStrAlign at https://huggingface.co/datasets/multimolecule/rnastralign (accessed on 15 January 2026); ArchiveII at https://huggingface.co/datasets/multimolecule/archiveii.1024 (accessed on 15 January 2026); bpRNA-1m at https://huggingface.co/datasets/multimolecule/bprna (accessed on 15 January 2026).

Conflicts of Interest

The authors declare no conflicts of interest.

Funding Statement

This work was supported by the National Natural Science Foundation of China under Grant 62271345, 62301356.

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.Spitale R.C., Incarnato D. Probing the dynamic RNA structurome and its functions. Nat. Rev. Genet. 2023;24:178–196. doi: 10.1038/s41576-022-00546-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Assmann S.M., Chou H.L., Bevilacqua P.C. Rock, scissors, paper: How RNA structure informs function. Plant Cell. 2023;35:1671–1707. doi: 10.1093/plcell/koad026. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Muñoz-Velasco I., Cruz-González A., Hernández-Morales R., Campillo-Balderas J.A., Cottom-Salas W., Jácome R., Vázquez-Salazar A. Pioneering role of RNA in the early evolution of life. Genet. Mol. Biol. 2024;47:e20240028. doi: 10.1590/1678-4685-gmb-2024-0028. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Mathews D.H., Moss W.N., Turner D.H. Folding and finding RNA secondary structure. Cold Spring Harb. Perspect. Biol. 2010;2:a003665. doi: 10.1101/cshperspect.a003665. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Ren J., Rastegari B., Condon A., Hoos H.H. HotKnots: Heuristic prediction of RNA secondary structures including pseudoknots. RNA. 2005;11:1494–1504. doi: 10.1261/rna.7284905. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Sato K., Kato Y., Hamada M., Akutsu T., Asai K. IPknot: Fast and accurate prediction of RNA secondary structures with pseudoknots using integer programming. Bioinformatics. 2011;27:i85–i93. doi: 10.1093/bioinformatics/btr215. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Fürtig B., Richter C., Wöhnert J., Schwalbe H. NMR spectroscopy of RNA. ChemBioChem. 2003;4:936–962. doi: 10.1002/cbic.200300700. [DOI] [PubMed] [Google Scholar]
  • 8.Fica S.M., Nagai K. Cryo-electron microscopy snapshots of the spliceosome: Structural insights into a dynamic ribonucleoprotein machine. Nat. Struct. Mol. Biol. 2017;24:791–799. doi: 10.1038/nsmb.3463. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Mathews D.H., Disney M.D., Childs J.L., Schroeder S.J., Zuker M., Turner D.H. Incorporating chemical modification constraints into a dynamic programming algorithm for prediction of RNA secondary structure. Proc. Natl. Acad. Sci. USA. 2004;101:7287–7292. doi: 10.1073/pnas.0401799101. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Turner D.H., Mathews D.H. NNDB: The nearest-neighbor parameter database for predicting stability of nucleic acid secondary structure. Nucleic Acids Res. 2010;38:D280–D282. doi: 10.1093/nar/gkp892. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Zuker M., Stiegler P. Optimal computer folding of large RNA sequences using thermodynamics and auxiliary information. Nucleic Acids Res. 1981;9:133–148. doi: 10.1093/nar/9.1.133. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Lorenz R., Bernhart S.H., Höner zu Siederdissen C., Tafer H., Flamm C., Stadler P.F., Hofacker I.L. ViennaRNA Package 2.0. Algorithms Mol. Biol. 2011;6:26. doi: 10.1186/1748-7188-6-26. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Zuker M. Mfold web server for nucleic acid folding and hybridization prediction. Nucleic Acids Res. 2003;31:3406–3415. doi: 10.1093/nar/gkg595. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Mathews D.H., Turner D.H. Prediction of RNA secondary structure by free energy minimization. Curr. Opin. Struct. Biol. 2006;16:270–278. doi: 10.1016/j.sbi.2006.05.010. [DOI] [PubMed] [Google Scholar]
  • 15.Bellaousov S., Mathews D.H. ProbKnot: Fast prediction of RNA secondary structure including pseudoknots. RNA. 2010;16:1870–1880. doi: 10.1261/rna.2125310. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Nawrocki E.P., Burge S.W., Bateman A., Daub J., Eberhardt R.Y., Eddy S.R., Floden E.W., Gardner P.P., Jones T.A., Tate J., et al. Rfam 12.0: Updates to the RNA families database. Nucleic Acids Res. 2015;43:D130–D137. doi: 10.1093/nar/gku1063. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Zhang T., Singh J., Litfin T., Zhan J., Paliwal K., Zhou Y. RNAcmap: A fully automatic pipeline for predicting contact maps of RNAs by evolutionary coupling analysis. Bioinformatics. 2021;37:3494–3500. doi: 10.1093/bioinformatics/btab391. [DOI] [PubMed] [Google Scholar]
  • 18.Dosovitskiy A., Beyer L., Kolesnikov A., Weissenborn D., Zhai X., Unterthiner T., Dehghani M., Minderer M., Heigold G., Gelly S., et al. An image is worth 16×16 words: Transformers for image recognition at scale. arXiv. 20202010.11929 [Google Scholar]
  • 19.Zhang Z., Xiao S., Lan G., Xi M., Wen J., Yang J. MAIP: A Multi-Attribute Informativeness Proxy for Image Semantic Broadcasting Communication. IEEE Trans. Broadcast. 2025;71:903–913. doi: 10.1109/TBC.2025.3573144. [DOI] [Google Scholar]
  • 20.Huo Z., Yang J., Chen D., Zhang L., Li Z., Sun L. Spatial interpolation of digital elevation model based on multi-scale conditional generative adversarial network with adaptive joint loss. J. Appl. Remote Sens. 2025;19:014504. doi: 10.1117/1.JRS.19.014504. [DOI] [Google Scholar]
  • 21.Chai X., Shao F., Chen H., Mu B., Ho Y.S. Super-Resolution Reconstruction for Stereoscopic Omnidirectional Display Systems via Dynamic Convolutions and Cross-View Transformer. IEEE Trans. Instrum. Meas. 2023;72:5025012. doi: 10.1109/TIM.2023.3309358. [DOI] [Google Scholar]
  • 22.Siddiqui F., Yang J., Xiao S., Fahad M. Diffusion model in modern detection: Advancing deepfake techniques. Knowl.-Based Syst. 2025;325:113922. doi: 10.1016/j.knosys.2025.113922. [DOI] [Google Scholar]
  • 23.Meng X., Wang N., Shao F., Li S. Vision transformer for pansharpening. IEEE Trans. Geosci. Remote Sens. 2022;60:5409011. doi: 10.1109/TGRS.2022.3168465. [DOI] [Google Scholar]
  • 24.Yang Y., Yi H., Xi M., Wen J., Yang J. GenAI-Driven Unsupervised Denoising for Consumer Device Imagery. IEEE Consum. Electron. Mag. 2025;14:94–102. doi: 10.1109/MCE.2025.3538904. [DOI] [Google Scholar]
  • 25.Xi M., Wen J., He J., Xiao S., Yang J. An expert experience-enhanced security control approach for AUVs of the underwater transportation cyber-physical systems. IEEE Trans. Intell. Transp. Syst. 2025;26:14086–14098. doi: 10.1109/TITS.2024.3524730. [DOI] [Google Scholar]
  • 26.Singh J., Hanson J., Paliwal K., Zhou Y. RNA secondary structure prediction using an ensemble of two-dimensional deep neural networks and transfer learning. Nat. Commun. 2019;10:5407. doi: 10.1038/s41467-019-13395-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Zhang H., Zhang C., Li Z., Li C., Wei X., Zhang B., Liu Y. A new method of RNA secondary structure prediction based on convolutional neural network and dynamic programming. Front. Genet. 2019;10:467. doi: 10.3389/fgene.2019.00467. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Fu L., Cao Y., Wu J., Peng Q., Nie Q., Xie X. UFold: Fast and accurate RNA secondary structure prediction with deep learning. Nucleic Acids Res. 2022;50:e14. doi: 10.1093/nar/gkab1074. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Chen X., Li Y., Umarov R., Gao X., Song L. RNA Secondary Structure Prediction By Learning Unrolled Algorithms; Proceedings of the International Conference on Learning Representations 2020; Virtual. 26 April–1 May 2020. [Google Scholar]
  • 30.Tan C., Gao Z., Gu L., Wang G., Wu L., Xia J., Zheng J., Li S.Z. RFold: RNA Secondary Structure Prediction with Decoupled Optimization; Proceedings of the International Conference on Learning Representations 2024; Vienna, Austria. 7–11 May 2024. [Google Scholar]
  • 31.Sato K., Akiyama M., Sakakibara Y. RNA secondary structure prediction using deep learning with thermodynamic integration. Nat. Commun. 2021;12:941. doi: 10.1038/s41467-021-21194-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Wang Y., Shen Y., Li J., Wang T., Peng J., Shang X. Enhanced RNA secondary structure prediction through integrative deep learning and structural context analysis. Nucleic Acids Res. 2025;53:gkaf533. doi: 10.1093/nar/gkaf533. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Jin K., Wang X., Shao F. Jointly Texture Enhanced and Stereo Captured Network for Stereo Image Super-Resolution. Pattern Recognit. Lett. 2023;167:141–148. doi: 10.1016/j.patrec.2023.02.015. [DOI] [Google Scholar]
  • 34.Hu J., Sun X., Bai X., Qin Y., Wang H., Han J. Subdivision Features-Guided Brain MRI Super-Resolution via Forward and Backward Propagation; Proceedings of the ICASSP 2024—2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); Seoul, Republic of Korea. 14–19 April 2024; pp. 1666–1670. [Google Scholar]
  • 35.Hu J., Qin Y., Wang H., Han J. MS2CAM: Multi-scale self-cross-attention mechanism-based MRI super-resolution. Displays. 2025;88:103033. doi: 10.1016/j.displa.2025.103033. [DOI] [Google Scholar]
  • 36.Wang Y., Liu Y., Wang S., Liu Z., Gao Y., Zhang H., Dong L. ATTfold: RNA secondary structure prediction with pseudoknots based on attention mechanism. Front. Genet. 2020;11:612086. doi: 10.3389/fgene.2020.612086. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Nussinov R., Pieczenik G., Griggs J.R., Kleitman D.J. Algorithms for loop matchings. SIAM J. Appl. Math. 1978;35:68–82. doi: 10.1137/0135006. [DOI] [Google Scholar]
  • 38.Bernhart S.H., Hofacker I.L., Will S., Gruber A.R., Stadler P.F. RNAalifold: Improved consensus structure prediction for RNA alignments. BMC Bioinform. 2008;9:474. doi: 10.1186/1471-2105-9-474. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Hamada M., Kiryu H., Sato K., Mituyama T., Asai K. Prediction of RNA secondary structure using generalized centroid estimators. Bioinformatics. 2009;25:465–473. doi: 10.1093/bioinformatics/btn601. [DOI] [PubMed] [Google Scholar]
  • 40.Akiyama M., Sato K., Sakakibara Y. A max-margin training of RNA secondary structure prediction integrated with the thermodynamic model. J. Bioinform. Comput. Biol. 2018;16:1840025. doi: 10.1142/S0219720018400255. [DOI] [PubMed] [Google Scholar]
  • 41.Tan Z., Fu Y., Sharma G., Mathews D.H. TurboFold II: RNA structural alignment and secondary structure prediction informed by multiple homologs. Nucleic Acids Res. 2017;45:11570–11581. doi: 10.1093/nar/gkx815. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Danaee P., Rouches M., Wiley M., Deng D., Huang L., Hendrix D. bpRNA: Large-scale automated annotation and analysis of RNA secondary structure. Nucleic Acids Res. 2018;46:5381–5394. doi: 10.1093/nar/gky285. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Sloma M.F., Mathews D.H. Exact calculation of loop formation probability identifies folding motifs in RNA secondary structures. RNA. 2016;22:1808–1818. doi: 10.1261/rna.053694.115. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Reuter J.S., Mathews D.H. RNAstructure: Software for RNA secondary structure prediction and analysis. BMC Bioinform. 2010;11:129. doi: 10.1186/1471-2105-11-129. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Do C.B., Woods D.A., Batzoglou S. CONTRAfold: RNA secondary structure prediction without physics-based models. Bioinformatics. 2006;22:e90–e98. doi: 10.1093/bioinformatics/btl246. [DOI] [PubMed] [Google Scholar]
  • 46.Huang L., Zhang H., Deng D., Zhao K., Liu K., Hendrix D.A., Mathews D.H. LinearFold: Linear-time approximate RNA folding by 5′-to-3′dynamic programming and beam search. Bioinformatics. 2019;35:i295–i304. doi: 10.1093/bioinformatics/btz375. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Qiu X. Sequence similarity governs generalizability of de novo deep learning models for RNA secondary structure prediction. PLoS Comput. Biol. 2023;19:e1011047. doi: 10.1371/journal.pcbi.1011047. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Zhao Q., Zhao Z., Fan X., Yuan Z., Mao Q., Yao Y. Review of machine learning methods for RNA secondary structure prediction. PLoS Comput. Biol. 2021;17:e1009291. doi: 10.1371/journal.pcbi.1009291. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Andronescu M., Aguirre-Hernandez R., Condon A., Hoos H.H. RNAsoft: A suite of RNA secondary structure prediction and design software tools. Nucleic Acids Res. 2003;31:3416–3422. doi: 10.1093/nar/gkg612. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Zakov S., Goldberg Y., Elhadad M., Ziv-Ukelson M. Rich parameterization improves RNA structure prediction. J. Comput. Biol. 2011;18:1525–1542. doi: 10.1089/cmb.2011.0184. [DOI] [PubMed] [Google Scholar]
  • 51.Jung A.J., Lee L.J., Gao A.J., Frey B.J. Rtfold: RNA secondary structure prediction using deep learning with domain inductive bias; Proceedings of the 2022 ICML Workshop on Computational Biology; Baltimore, MD, USA. 17–23 July 2022. [Google Scholar]
  • 52.Bugnon L.A., Di Persia L., Gerard M., Raad J., Prochetto S., Fenoy E., Chorostecki U., Ariel F., Stegmayer G., Milone D.H. sincFold: End-to-end learning of short-and long-range interactions in RNA secondary structure. Briefings Bioinform. 2024;25:bbae271. doi: 10.1093/bib/bbae271. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Zhang H., Zhang L., Mathews D.H., Huang L. LinearPartition: Linear-time approximation of RNA folding partition function and base-pairing probabilities. Bioinformatics. 2020;36:i258–i267. doi: 10.1093/bioinformatics/btaa460. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Kothapalli A., Setty S. Cyclefold: Folding-scheme-based recursive arguments over a cycle of elliptic curves. Cryptol. ePrint Arch. 2023 [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The data were derived from the following resources available in the public domain: RNAStrAlign at https://huggingface.co/datasets/multimolecule/rnastralign (accessed on 15 January 2026); ArchiveII at https://huggingface.co/datasets/multimolecule/archiveii.1024 (accessed on 15 January 2026); bpRNA-1m at https://huggingface.co/datasets/multimolecule/bprna (accessed on 15 January 2026).


Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES