Skip to main content
Frontiers in Bioengineering and Biotechnology logoLink to Frontiers in Bioengineering and Biotechnology
. 2026 Mar 19;14:1762967. doi: 10.3389/fbioe.2026.1762967

Light-RepViTSR: ultra-lightweight super-resolution for real-time photoacoustic endoscopy in tumor biopsy

Guanyi Jiang 1, Rui Yang 2, Zhanfeng Fang 2, Yuwei Luo 2, Xianghu Yu 2, Li Liu 2,*
PMCID: PMC13044124  PMID: 41938057

Abstract

Real-time in situ biopsy marks a paradigm shift in clinical oncology by enabling immediate intraprocedural pathological diagnosis during endoscopy. Photoacoustic endoscopy (PAE) is a pivotal technology in this field that uniquely visualizes tumor microvasculature and hypoxia through the synergistic fusion of optical contrast and ultrasonic resolution. However, PAE’s inherent resolution-speed tradeoff in raster scanning induces severe motion artifacts from physiological activities (e.g., peristalsis and respiration), critically compromising diagnostic reliability. Although deep-learning-based super-resolution (SR) techniques show promise for photoacoustic microscopy, their clinical translation to PAE is hindered by excessive computational demands and insufficient real-time performance. To overcome this limitation, we propose Light-RepViTSR, an ultra-lightweight SR reconstruction network based on the RepViT architecture and specifically optimized for real-time PAE. Our approach integrates the representational capacity of RepViT’s re-parameterizable convolutional blocks while eliminating non-essential components (e.g., squeeze-and-excitation layers) to maximize computational efficiency. Comprehensive evaluation on a multi-source dataset—including 19 previously unseen murine cerebrovascular images and 18 self-collected plant vein images—demonstrates the superiority of Light-RepViTSR. The network consistently outperforms conventional methods across scaling factors ( ×2 , ×4 , and ×8 ) to achieve significant improvements in PSNR (up to +1.41  dB at ×8 ) and SSIM (up to +0.047 at ×2 ) while reducing model size by >99% and inference time by >60% versus SRResNet. This study establishes a pathway toward practical real-time high-resolution PAE, demonstrating significant potential for enhancing in situ tumor biopsy accuracy.

Keywords: deep-learning, in situ biopsy, lightweight neural network, photoacoustic endoscopy, real-time imaging, RepVit, super-resolution

1. Introduction

Early and accurate malignancy diagnosis is fundamental to successful cancer management (Roukos et al., 2007; Schiffman et al., 2015; Das et al., 2023). In situ biopsy during endoscopic procedures offers transformative potential by enabling immediate histopathological assessment without repeated invasive interventions (Krafft and Popp, 2023). Photoacoustic imaging (PAI), a hybrid modality that leverages the photoacoustic effect, has emerged as a powerful tool for visualizing functional and molecular biomarkers in tumor biology (Lin and Wang, 2022; Liu et al., 2019). This effect involves pulsed laser absorption by chromophores (e.g., hemoglobin and melanin) to generate thermoelastic expansion that emits broadband ultrasonic waves detected for image formation (Figure 1a).

FIGURE 1.

Composite scientific illustration showing: (a) Photoacoustic effect: pulsed laser on tissue chromophores induces thermoelastic expansion and ultrasonic wave generation; (b) Schematic depicting motion artifacts in PAE caused by slow raster scanning, where physiological tissue movement (e.g., peristalsis, respiration) between scan lines results in image misalignment, blurring, and structural distortion of the acquired photoacoustic images.

(a) Photoacoustic effect schematic: pulsed laser absorption by tissue chromophores generates ultrasonic waves that aredetected using an acoustic transducer. (b) Motion artifacts in PAE slow raster scanning cause image misalignment and distortion due to tissue movement between scan lines.

Photoacoustic microscopy (PAM) provides high-resolution microvasculature imaging in preclinical and clinical contexts (Hu and Wang, 2010; Mirg et al., 2022). Crucially, PAI miniaturization into photoacoustic endoscopy (PAE) extends this capability to intraluminal imaging (Yoon and Cho, 2013). By integrating miniature scanning units into endoscopic probes, PAE visualizes subsurface microvasculature and hypoxia within gastrointestinal, respiratory, and other luminal organs (Song et al., 2025; Park et al., 2025). This capability is paramount for in situ biopsy, enabling lesion targeting based on malignancy hallmarks such as angiogenesis and altered oxygen metabolism. The real-time acquisition of optically stained histological data during examinations could substantially reduce diagnostic delays between lesion identification, sampling, and pathological verification (Ciepła and Smolarczyk, 2024; Majidpoor and Mortezaee, 2021).

Despite its promise, PAE inherits a critical limitation from high-resolution PAM systems: the point-scanning mechanism necessitates a fundamental trade-off between spatial resolution and imaging speed. High-resolution imaging requires dense spatial sampling, prolonging acquisition time (Kaur et al., 2020; Wu et al., 2024). This slow acquisition renders systems vulnerable to motion artifacts from physiological processes that include visceral peristalsis, cardiac pulsation, and respiration (Mikami et al., 2016; Rodrigues et al., 2020). As shown in Figure 1b, such motions induce severe tissue deformation and inter-scanline blurring, significantly degrading image quality and potentially causing misdiagnosis or missed pathological features (Hussain et al., 2022; Baad et al., 2017). Consequently, motion artifact challenges in PAM are exacerbated in PAE’s confined, dynamic environment. An effective mitigation strategy involves accelerated acquisition through reduced sampling points, followed by super-resolution (SR) algorithms to reconstruct anatomical structures from under-sampled data while preserving essential information for clinical real-time imaging (Qiu et al., 2023; Shin et al., 2024; Zhao et al., 2019).

Deep-learning-based SR techniques have been extensively investigated for medical image enhancement (Pain et al., 2022; Bashir et al., 2021; Yang et al., 2023). While several SR networks reconstruct high-resolution (HR) PAM images from low-resolution (LR) inputs, they typically target offline processing with substantial computational complexity and model sizes, thus failing to meet the stringent latency and hardware constraints of real-time PAE.

Recent advances in lightweight vision architectures demonstrate that convolutional networks can achieve state-of-the-art performance with minimal computational overhead (Chen et al., 2024; Cong and Zhou, 2023; Liu et al., 2024). RepViT (Wang et al., 2024) notably reformulates mobile CNN design using vision transformer principles, demonstrating that re-parameterizable convolutional blocks outperform both lightweight CNNs and vision transformers on mobile platforms. Its key innovation involves separating token and channel mixing operations while maintaining convolutional hardware efficiency through structural re-parameterization (Zhang et al., 2025; Guo et al., 2025).

Building upon these advances, we introduce Light-RepViTSR, an ultra-lightweight SR network specifically optimized for real-time PAE through RepViT adaptation. Our contributions are threefold.

  1. A novel SR architecture that adapts RepViT’s efficient design principles and is optimized for PAE by employing identity mapping blocks that exclude computationally intensive squeeze-and-excitation layers.

  2. A rigorous evaluation framework using unseen test data (19 murine cerebrovascular and 18 plant vein images) to assess generalization capability and clinical relevance.

  3. Comprehensive quantitative and qualitative analyses that demonstrate Light-RepViTSR’s superior SR performance versus state-of-the-art methods, with orders-of-magnitude fewer parameters and significantly reduced inference time, establishing a new benchmark for practical SR in real-time PAE.

2. Methods

2.1. Network architecture

Light-RepViTSR’s design optimizes the balance between representational capacity and computational efficiency for real-time photoacoustic endoscopy. As depicted in Figure 2, the architecture comprises four core components: stem convolution, identity-mapping RepViT blocks for deep feature extraction, residual connection, and efficient up-sampling.

FIGURE 2.

Diagram of a deep learning model for image processing showing a low-resolution 64 by 64 grayscale image passing through sequential modules labeled Stem, Light-RepViT blocks, convolution, PixelShuffler, additional convolution, and Tanh activation, resulting in a higher resolution 256 by 256 grayscale output.

Light-RepViTSR architecture highlighting the stem convolution, RepViT block sequence, residual connection, and up-sampling module.

2.1.1. Stem module

The input low-resolution image ILRRH×W×3 undergoes initial processing through a 3×3 convolutional layer with 16 output channels, yielding feature representations F0RH×W×16 . This minimalist stem design minimizes computational overhead at the highest spatial resolution while preserving adequate feature extraction capability. We employ Kaiming normal initialization with ReLU nonlinearity to ensure stable gradient propagation during training.

2.1.2. Deep feature extraction

Shallow features F0 that constitute the network’s core innovation are processed through three cascaded RepViT blocks configured for identity mapping. Diverging from the original RepViT design (Wang et al., 2023), we deliberately exclude squeeze-and-excitation (SE) layers based on ablation studies that demonstrate negligible performance gains for super-resolution tasks relative to their computational cost. Each block implements efficient separation of token and channel mixing operations while leveraging structural re-parameterization for inference-time efficiency.

2.1.3. Residual connection

A global residual connection combines the original stem features F0 with the RepViT block sequence output. This configuration preserves essential low-frequency information and enhances gradient flow during optimization—particularly critical for stabilizing deep network training. The residual pathway enables the network to focus on learning high-frequency components required for detail restoration rather than full-image reconstruction.

2.1.4. Efficient up-sampling

Enhanced features are transformed to the target resolution through a unified up-sampling module. For all scaling factors ( ×2 , ×4 , and ×8 ), we implement (1) a 3×3 convolution expanding channel dimensionality, (2) pixel-shuffle spatial rearrangement, and (3) a final 3×3 convolution to produce the super-resolved output ISRRsH×sW×3 ( s = scaling factor). This approach maintains reconstruction quality while avoiding computationally expensive transposed convolutions.

2.1.5. Output activation

A tanh activation constrains final outputs to [1,1] , enhancing training stability and ensuring physically plausible intensity ranges. During inference, outputs are linearly rescaled to standard [0,255] for visualization and analysis.

2.2. RepViT block design

The RepViT block (Figure 3) constitutes the fundamental building block of Light-RepViTSR. We exclusively employ the identity mapping variant, mathematically formalized as follows.

FIGURE 3.

Diagram of the Light-RepViT Block showing two main sections: Token Mixer on the left with parallel 3x3DW and 1x1DW convolution and batch normalization, and Channel Mixer on the right with sequential 1x1 convolution, batch normalization, GELU activation, and a second 1x1 convolution and batch normalization, with arrows indicating data flow.

RepViT block structure featuring identity mapping with separate token and channel mixers and SE-layer omission for enhanced efficiency.

Token mixer: it utilizes a RepVGGDW module comprising three parallel branches during training:

  • 3×3 depthwise convolution with batch normalization;

  • 1×1 depthwise convolution with batch normalization; and

  • identity branch.

Output aggregation follows Equation 1:

TokenMixerx=DWConv3×3x+DWConv1×1x+x. (1)

During inference, structural re-parameterization consolidates these branches into a single 3×3 depthwise convolution that maintains functionality while substantially reducing computational complexity and memory access costs.

Channel mixer: it implements a transformer-inspired two-layer MLP with expansion ratio using two Equation 2:

ChannelMixerx=x+Conv1×1GELUConv1×1x. (2)

The first convolution expands channels from C to 2C , while the second projects back to C . GELU activation provides smooth nonlinear transformation with favorable optimization properties.

Design rationale: comprehensive ablation studies motivated the deliberate exclusion of SE layers present in some RepViT variants. While SE layers enhance classification performance through channel-dependency modeling, they yield diminishing returns for super-resolution applications while increasing computational overhead. The identity mapping configuration provides sufficient representational capacity for capturing the local and global dependencies essential to high-quality reconstruction while maximizing computational efficiency—a critical requirement for real-time PAE.

2.3. Data preparation and experimental setup

We curated a multi-source dataset encompassing diverse biological structures and imaging conditions to ensure model robustness and generalizability.

Data sources.

  1. Public Murine Cerebrovascular Dataset (Duke PAM): a high-resolution photoacoustic microscopy dataset of mouse brain vasculature (Chen et al., 2019) providing pristine, high-SNR images with intricate vascular networks. These served as high-resolution ground truth for direct comparison with existing literature.

  2. Custom plant vein dataset: novel leaf vein images acquired using our in-house optical-resolution PAM system (Figure 4a). The fractal-like branching patterns of plant veins exhibit topological similarities to tumor microvasculature and offer a cost-effective model for assessing generalization to unseen biological structures. Representative images appear in Figure 4b.

FIGURE 4.

Panel (a) shows a schematic diagram of a photoacoustic imaging system using a 532 nanometer laser, optical components, scanning stage, data acquisition, and computer for imaging a leaf sample. Panel (b) presents a color-coded photoacoustic intensity image of a leaf vein network, with intensity ranging from minimum (dark red) to maximum (bright yellow) as indicated by the adjacent scale.

(a) Custom-built optical-resolution photoacoustic microscopy platform for plant vein acquisition. (b) Representative maximum amplitude projection (MAP) images demonstrating complex branching patterns used for generalization testing. Scale bar: 500 μ m.

PAM system parameters. For our custom-built PAM system used to acquire plant vein data, the specific parameters are as follows. (1) Optical parameters: laser wavelength (532 nm), pulse duration (15 ns), and pulse energy (0.0637 mJ/cm2). (2) Acoustic parameters: ultrasound transducer frequency (50 MHz), bandwidth (50%), and sampling rate (200 MHz). (3) PAM system specifications: scanning mode (raster scan) and step size (10 μ  m).

Data preparation. From the Duke dataset, we randomly extracted 5,000 non-overlapping 256×256 patches for training and 500 for validation. Testing employed 19 completely unseen Duke images and 18 plant vein images to evaluate cross-domain generalization. All images were changed to [1,1] for training stability.

PAE under-sampling simulation. To simulate low-resolution inputs from accelerated PAE scanning, we applied raster under-sampling to high-resolution ground truth (Figure 5). For scaling factor s , we retained every s -th row and column, reducing sampling density by s2 . This mimics practical PAE scenarios where increased speed compromises spatial resolution. The resulting sparse images were interpolated to original 256×256 resolution using our model, with s={2,4,8} evaluating performance across under-sampling severities.

FIGURE 5.

Three side-by-side diagrams show grid patterns with colored highlights: yellow on the left, green in the center, and red on the right. Below each, arrows point downward to smaller colored grids with multipliers indicating how many sub-grids can be formed: two yellow, four green, and eight red. The illustration visually represents hierarchical or recursive division of grids into smaller colored blocks.

Raster under-sampling procedure for scaling factors ×2 , ×4 , and ×8 .

2.4. Data processing workflow

Figure 6 illustrates the complete pipeline from raw data acquisition to super-resolution reconstruction and quantitative evaluation. The process consists of the following key stages. (1) High-resolution (HR) ground truth: the pipeline begins with high-resolution PA images, which serve as the reference standard and are used as training labels for model supervision. (2) Low-resolution (LR) image generation via down-sampling: HR images undergo raster under-sampling ( × 4 scale illustrated) to simulate the low-resolution, blurred inputs resulting from accelerated PAE scanning. This step mimics the trade-off between imaging speed and spatial resolution in clinical settings. (3) Model training and SR reconstruction: the LR images are fed into the Light-RepViTSR model during training, where the network learns the mapping from LR to HR representations. During inference, the trained model reconstructs super-resolved (SR) images from LR inputs. (4) Quantitative evaluation: the reconstructed SR images are compared against the original HR ground truth using three established metrics: root mean square error (RMSE, lower values indicate less error), peak signal-to-noise ratio (PSNR, higher values indicate better fidelity), and structural similarity index (SSIM, closer to 1 indicates better structural preservation). This workflow provides a clear, closed-loop representation of the SR task, demonstrating how data flows through generation, processing, reconstruction, and validation phases.

FIGURE 6.

Workflow diagram showing high-resolution (HR) images being downsampled to create low-resolution (LR) images, which then enter a model training box. The model generates super-resolution (SR) images, which are evaluated using RMSE, PSNR, and SSIM metrics.

Data processing workflow for photoacoustic super-resolution reconstruction.

2.5. Evaluation metrics

We employed three full-reference image quality metrics to quantitatively assess super-resolution performance.

  • Root mean square error: it quantifies pixel-wise intensity differences between reconstructed and ground truth images, where lower values indicate superior fidelity Equation 3:

RMSE=1Ni=1NYiY^i2, (3)

with Yi representing the ground truth pixel value, Y^i representing the reconstructed pixel value, and N representing the total pixels.

  • Peak signal-to-noise ratio: this is a logarithmic fidelity metric derived from mean squared error, where higher values denote improved reconstruction quality Equation 4:

PSNR=20log10MAXIRMSE, (4)

where MAXI represents the maximum pixel value (1.0 for normalized images).

  • Structural similarity index: this is a perceptual metric evaluating luminance, contrast, and structural fidelity, ranging from −1 to 1 (1 indicates perfect similarity) Equation 5:

SSIMx,y=2μxμy+C12σxy+C2μx2+μy2+C1σx2+σy2+C2, (5)

where μ represents the local mean, σ represents the standard deviation, σxy represents the cross-covariance, and C represents the stabilization constants.

2.6. Loss function and optimization strategy

Loss function. We implemented StableMSELoss to mitigate numerical instability while preserving mean squared error benefits Equation 6:

LStableMSE=1Ni=1NYiY^i2+ϵ, (6)

where ϵ=1×108 prevents numerical underflow. This stabilization proved critical during early training stages and for challenging ×8 super-resolution.

2.6.1. Optimization

We employed AdamW with hyperparameters:

  • learning rate: 1×104

  • betas: (0.9, 0.999)

  • epsilon: 1×108

  • weight decay: 1×104

Training spanned 130 epochs (batch size = 10) with cosine annealing scheduler (Tmax=130) . This configuration ensured robust convergence and prevented overfitting, which was particularly valuable, given medical imaging’s typically limited datasets.

2.7. Implementation details

2.7.1. Training strategy

We trained specialized models per scaling factor ( ×2 , ×4 , and ×8 ):

  • ×2/×4 scaling: patch-based training using randomly cropped 128×128 patches from 256×256 images, subjected to raster under-sampling and bicubic down-sampling. This approach (1) augmented data diversity, reducing overfitting; (2) focused learning on local structural patterns critical for SR; and (3) reduced memory demands, enabling larger batches.

  • ×8 scaling: full 256×256 image training to provide global contextual information essential for reconstructing plausible anatomical structures amidst extreme information loss. This prevents compounding artifacts from under-sampling and cropping.

Consistent framework. All models trained for 130 epochs (batch size = 10) using the AdamW optimizer and learning rate scheduler from Section 2.6 to ensure fair comparison while addressing scaling-specific requirements. Fair comparison protocol: all models were trained using identical training, validation, and test sets. Baseline models (SRResNet, SRGAN (Ledig et al., 2017), EDSR (Lim et al., 2017), DDBPN (Haris et al., 2018), and MobileSR (Zhang et al., 2020)) were trained using their respective recommended optimizers, loss functions, learning rates, and scheduling strategies from their original studies, and we selected their best-performing results for evaluation. The patch-based training for × 2/ × 4 and full-image training for × 8 was applied to all methods compared to ensure consistent conditions for evaluating their ability to handle different levels of information loss.

2.8. Experimental protocol

Quantitative results report a mean ± standard deviation across test sets. For inference time measurements, we implemented a rigorous protocol, excluding first-image inference times to account for GPU warm-up effects while reporting statistics only on subsequent images to reflect sustained performance.

2.9. Statistical analysis

All comparisons employed rigorous statistical testing using Shapiro–Wilk normality tests (α=0.05) , followed by Wilcoxon signed-rank tests for non-normal distributions. Effect sizes were calculated using Cohen’s d. All reported differences are statistically significant at α=0.05 unless otherwise noted.

2.9.1. Platform

Experiments were conducted on an Intel i7-14700K CPU, 64 GB RAM, and NVIDIA GeForce RTX 4070 GPU (12 GB VRAM) using PyTorch 2.8.0 and CUDA 12.6.

3. Results and analysis

3.1. Qualitative reconstruction performance

Visual comparisons on the Duke murine cerebrovascular dataset (Figure 7) demonstrate the superior capability of Light-RepViTSR in reconstructing fine anatomical details from severely under-sampled inputs. Across all scaling factors ( ×2 , ×4 , and ×8 ), Light-RepViTSR consistently recovers finer vascular structures and sharper edges than conventional SR methods including SRResNet, SRGAN, EDSR, DDBPN, and MobileSR.

FIGURE 7.

Comparison graphic of high-resolution (HR), undersampled, and six super-resolution imaging methods (SRResNet, SRGAN, EDSR, DDBPN, MobileSR, Light-RepViTSR) applied at three sampling rates (×2, ×4, ×8). Images show vascular structures with normalized photoacoustic (PA) intensity. Red boxes highlight regions of interest. Each method includes peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) values for quantitative comparison. A grayscale bar indicates PA intensity from minimum to maximum.

Super-resolution reconstruction of murine cerebrovascular structures across scaling factors ( ×2 , ×4 , and ×8 ). Red boxes highlight detailed regions for visual comparison. Light-RepViTSR demonstrates superior recovery of fine vascular details, structural continuity, and boundary sharpness compared to conventional SR methods, particularly at high scaling factors where information loss is most severe. Scale bars: 500 μ  m.

The qualitative analysis reveals several key observations:

  • Structural continuity: at ×8 scaling, where information loss is most severe, Light-RepViTSR maintains superior structural continuity in capillary networks, thus minimizing fragmentation and disconnections commonly observed in other methods.

  • Boundary sharpness: vessel boundaries reconstructed by Light-RepViTSR exhibit significantly enhanced sharpness, particularly in thin vascular structures that are critical for accurate microvasculature analysis.

  • Noise suppression: compared to MobileSR and other lightweight architectures, Light-RepViTSR demonstrates improved noise control, thus generating cleaner reconstructions with reduced artifactual speckling.

  • Detail preservation: fine branching patterns and intricate vascular connections are better preserved across all scaling factors with minimal loss of morphological information.

These advantages extend to the plant vein dataset (Figure 8), where Light-RepViTSR demonstrates robust generalization to diverse biological structures with distinct topological characteristics.

FIGURE 8.

Cross-domain super-resolution reconstruction comparison on plant vein vascular networks, with three rows for ×2, ×4, ×8 undersampling rates. Columns show original high-resolution (HR) reference, undersampled images, and reconstructions from six methods (SRResNet, SRGAN, EDSR, DDBPN, MobileSR, Light-RepViTSR). Each panel includes PSNR/SSIM quantitative metrics; red boxes highlight key regions for visual detail comparison. A right-side grayscale bar indicates normalized photoacoustic (PA) intensity from minimum to maximum.

Cross-domain generalization on plant vein structures across scaling factors ( ×2 , ×4 , and ×8 ). Red boxes highlight detailed regions showing complex branching patterns. Light-RepViTSR accurately reconstructs hierarchical vein organization with preserved connectivity, minimal topological distortion, and natural diameter variations, demonstrating robust generalization to diverse biological structures. Scale bars: 1 mm .

The plant vein reconstructions highlight Light-RepViTSR’s ability to achieve the following.

  • Maintain fractal-like branching: complex fractal branching patterns characteristic of plant vasculature are faithfully reconstructed with minimal topological distortion.

  • Preserve diameter variations: natural variations in vein diameter are maintained across scaling factors, avoiding the uniform thickening or thinning observed in other methods.

  • Ensure global connectivity: overall vascular network connectivity is preserved, which is crucial for accurate morphological analysis in biomedical applications.

3.2. Quantitative evaluation with statistical significance

Duke Cerebrovascular Dataset (19 images). Table 1 presents comprehensive quantitative results with rigorous statistical analysis. Light-RepViTSR demonstrates significant performance advantages across all scaling factors, with particularly pronounced improvements at higher scaling factors.

TABLE 1.

Quantitative evaluation with statistical significance on Duke murine cerebrovascular test set (19 images; mean ± SD).

Model Size (MB) Time (s) RMSE PSNR (dB) SSIM
SRResNet ×2 1.402 0.0124 ± 0.0021 5.20 ± 1.31 28.50 ± 4.27 0.840 ± 0.043
SRGAN ×2 1.406 0.0142 ± 0.0045 7.07 ± 1.34 24.16 ± 5.05 0.621 ± 0.070
EDSR ×2 1.398 0.0090 ± 0.0012 5.04 ± 1.32 29.10 ± 4.22 0.874 ± 0.041
DDBPN ×2 1.405 0.0100 ± 0.0041 5.01 ± 1.31 29.38 ± 4.13 0.873 ± 0.037
MobileSR ×2 0.895 0.0105 ± 0.0014 5.03 ± 1.32 29.62 ± 4.11 0.884 ± 0.036
Light-RepViTSR ×2 0.014 0.0046 ± 0.0045 5.08 ± 1.35 29.51 ± 4.11 0.886 ± 0.041
SRResNet ×4 1.549 0.0107 ± 0.0042 7.04 ± 1.29 20.52 ± 3.41 0.552 ± 0.101
SRGAN ×4 1.553 0.0111 ± 0.0048 8.29 ± 0.91 20.62 ± 3.32 0.377 ± 0.061
EDSR ×4 1.545 0.0063 ± 0.0025 6.94 ± 1.29 21.43 ± 3.47 0.588 ± 0.094
DDBPN ×4 1.552 0.0118 ± 0.0031 6.91 ± 1.30 21.81 ± 3.49 0.596 ± 0.093
MobileSR ×4 1.142 0.0112 ± 0.0033 6.87 ± 1.32 22.08 ± 3.54 0.618 ± 0.096
Light-RepViTSR ×4 0.042 0.0067 ± 0.0027 6.86 ± 1.32 22.03 ± 3.56 0.616 ± 0.095
SRResNet ×8 1.697 0.0116 ± 0.0046 7.71 ± 1.26 16.46 ± 3.34 0.383 ± 0.129
SRGAN ×8 1.701 0.0138 ± 0.0049 8.13 ± 0.86 17.47 ± 3.27 0.297 ± 0.057
EDSR ×8 1.693 0.0063 ± 0.0026 7.71 ± 1.28 16.90 ± 3.44 0.378 ± 0.125
DDBPN ×8 1.700 0.0099 ± 0.0038 7.70 ± 1.26 17.34 ± 3.40 0.393 ± 0.122
MobileSR ×8 1.390 0.0094 ± 0.0025 7.72 ± 1.26 17.56 ± 3.46 0.404 ± 0.123
Light-RepViTSR ×8 0.153 0.0044 ± 0.0036 7.68 ± 1.28 17.87 ± 3.60 0.419 ± 0.126

Bold values indicate the optimal performance metrics for each scaling factor (×2/×4/×8) on the respective datasets: minimum Size(MB), minimum Time(s), minimum RMSE, maximum PSNR, and maximum SSIM across all compared super-resolution models.

×2 scaling: Light-RepViTSR achieves a PSNR of 29.51 dB, significantly outperforming SRResNet (28.50 dB, p < 0.00001, Cohen’s d = 0.24) and SRGAN (24.16 dB, p < 0.00001, Cohen’s d = 1.16). Compared to MobileSR (29.62 dB), Light-RepViTSR shows statistically comparable performance (p = 0.00013, Cohen’s d = −0.027) while achieving 56.3% faster inference time (0.0046 s vs. 0.0105 s) and 98.4% smaller model size (0.014 MB vs. 0.895 MB).

×4 scaling: Light-RepViTSR maintains competitive performance with PSNR of 22.03 dB, significantly better than SRResNet (20.52 dB, p < 0.00001, Cohen’s d = 0.43) and SRGAN (20.62 dB, p < 0.00001, Cohen’s d = 0.41). Although slightly lower than MobileSR (22.08 dB, p = 0.0028, Cohen’s d = −0.013), the performance difference is negligible (0.05 dB), while Light-RepViTSR provides 40.6% faster inference and a 96.3% smaller model.

×8 scaling: Light-RepViTSR excels at the most challenging scaling factor, achieving a PSNR of 17.87 dB—significantly superior to all competing methods (all p-values < 0.05). It outperforms MobileSR by 0.31 dB (p < 0.00001, Cohen’s d = 0.089), DDBPN by 0.53 dB (p < 0.00001, Cohen’s d = 0.15), and SRResNet by 1.41 dB (p < 0.00001, Cohen’s d = 0.41). This demonstrates Light-RepViTSR’s exceptional capability in reconstructing plausible anatomical structures from extreme under-sampling.

Plant vein dataset (18 images): Table 2 confirms robust cross-domain performance. Light-RepViTSR achieves competitive reconstruction quality while maintaining superior efficiency.

  • ×2 scaling: comparable PSNR to DDBPN (29.64 dB vs. 29.64 dB, p = 0.67) while providing 34.1% faster inference.

  • ×4 scaling: competitive performance (PSNR: 24.23 dB) with 35.7% faster inference than DDBPN.

  • ×8 scaling: outperforms SRResNet by 0.26 dB (p < 0.00001) with 36.6% faster inference.

TABLE 2.

Quantitative evaluation with statistical significance on plant vein test set (18 images; mean ± SD).

Model Size (MB) Time (s) RMSE PSNR (dB) SSIM
SRResNet ×2 1.402 0.0123 ± 0.0049 6.65 ± 0.70 29.18 ± 2.31 0.744 ± 0.079
SRGAN ×2 1.406 0.0071 ± 0.0036 9.17 ± 0.83 19.64 ± 4.41 0.487 ± 0.198
EDSR ×2 1.398 0.0102 ± 0.0024 6.59 ± 0.71 29.53 ± 2.20 0.757 ± 0.079
DDBPN ×2 1.405 0.0132 ± 0.0035 6.55 ± 0.70 29.64 ± 2.18 0.762 ± 0.077
MobileSR ×2 0.895 0.0106 ± 0.0028 6.49 ± 0.67 29.80 ± 2.04 0.768 ± 0.076
Light-RepViTSR ×2 0.014 0.0087 ± 0.0048 6.57 ± 0.70 29.64 ± 2.15 0.762 ± 0.079
SRResNet ×4 1.549 0.0128 ± 0.0058 8.17 ± 0.54 24.02 ± 1.24 0.659 ± 0.078
SRGAN ×4 1.553 0.0116 ± 0.0041 8.72 ± 0.53 22.52 ± 2.01 0.554 ± 0.114
EDSR ×4 1.545 0.0077 ± 0.0035 8.16 ± 0.58 24.25 ± 1.17 0.670 ± 0.076
DDBPN ×4 1.552 0.0123 ± 0.0054 8.16 ± 0.59 24.25 ± 1.16 0.671 ± 0.076
MobileSR ×4 1.142 0.0117 ± 0.0037 8.20 ± 0.55 24.21 ± 1.14 0.664 ± 0.074
Light-RepViTSR ×4 0.042 0.0079 ± 0.0044 8.18 ± 0.56 24.23 ± 1.14 0.669 ± 0.074
SRResNet ×8 1.697 0.0112 ± 0.0071 9.12 ± 0.40 17.79 ± 0.75 0.489 ± 0.047
SRGAN ×8 1.701 0.0110 ± 0.0060 9.52 ± 0.29 16.84 ± 1.03 0.404 ± 0.058
EDSR ×8 1.693 0.0077 ± 0.0025 9.07 ± 0.45 17.87 ± 0.76 0.502 ± 0.049
DDBPN ×8 1.700 0.0123 ± 0.0068 9.04 ± 0.46 18.00 ± 0.77 0.509 ± 0.050
MobileSR ×8 1.390 0.0102 ± 0.0027 9.10 ± 0.42 17.95 ± 0.76 0.502 ± 0.047
Light-RepViTSR ×8 0.153 0.0071 ± 0.0036 9.12 ± 0.41 18.04 ± 0.77 0.504 ± 0.046

Bold values indicate the optimal performance metrics for each scaling factor (×2/×4/×8) on the respective datasets: minimum Size(MB), minimum Time(s), minimum RMSE, maximum PSNR, and maximum SSIM across all compared super-resolution models.

Key performance advantages.

  • Superior high-scaling performance: best performance at ×8 scaling, outperforming all competing methods with statistical significance.

  • Exceptional efficiency: 98.4% model size reduction and 34%–56% inference acceleration compared to some methods.

  • Strong generalization: consistent performance across in-domain (cerebrovascular) and out-of-domain (plant vein) datasets.

  • Optimal balance: best trade-off between reconstruction quality and computational efficiency for real-time applications.

3.3. Ablation study

Table 3 summarizes the ablation studies ( ×4 SR, Duke dataset) that validate our architectural design decisions. The results demonstrate that Light-RepViTSR achieves the optimal balance between performance and efficiency.

TABLE 3.

Ablation study ( ×4 SR, Duke test set; mean ± SD).

Configuration Size (MB) Time (s) RMSE PSNR (dB) SSIM
No residual 0.042 0.00312 ± 0.0006 6.87 ± 1.33 22.03 ± 3.60 0.620 ± 0.097
With SE 0.043 0.00300 ± 0.0006 6.90 ± 1.32 21.83 ± 3.63 0.612 ± 0.097
Light-RepViTSR 0.042 0.00277 ± 0.0005 6.88 ± 1.32 22.03 ± 3.56 0.616 ± 0.095

Residual connection impact: removing residual connections increases inference latency by 12.3% while providing minimal performance benefits, thus confirming the importance of residual connections for maintaining computational efficiency without compromising reconstruction quality.

SE layer exclusion: incorporating SE layers leads to performance degradation (−0.20 dB PSNR) with increased latency (8.2%) and model size (2.4%), empirically justifying their omission for super-resolution tasks where channel-dependency modeling yields diminishing returns.

Optimal configuration: the final Light-RepViTSR architecture achieves the best performance–efficiency balance, with minimal latency (0.00277 s) and model size (0.042 MB) while maintaining competitive reconstruction quality (PSNR: 22.03 dB, SSIM: 0.616).

4. Discussion

Our comprehensive evaluation establishes Light-RepViTSR as a state-of-the-art solution for efficient, high-quality super-resolution in photoacoustic endoscopy. The network’s exceptional capability to reconstruct fine microvascular details from severely under-sampled inputs—validated by statistically significant improvements on unseen test data—directly addresses the persistent motion artifact challenge inherent in slow PAE scanning protocols.

4.1. Performance superiority

Light-RepViTSR demonstrates statistically significant advantages over traditional SR methods, with PSNR improvements up to +1.41 dB at ×8 scaling (p < 0.00001, Cohen’s d = 0.41). Notably, it achieves its most significant advantage at the most challenging scaling factor, where information loss is greatest. This capability is particularly valuable for PAE applications, where high-speed imaging necessitates severe under-sampling. The network’s ability to maintain structural continuity and boundary sharpness at high scaling factors represents a substantial advance over existing methods.

4.2. Computational efficiency

Light-RepViTSR’s core innovation lies in its strategic adaptation of RepViT principles. By maintaining the representational capacity of re-parameterizable convolutional blocks while eliminating non-essential components (e.g., SE layers), the network achieves unprecedented efficiency: model sizes reduced by 98.4% and inference times accelerated by 34%–56% compared to some methods. This efficiency enables real-time processing on commodity hardware, fulfilling critical clinical deployment requirements for endoscopic applications where computational resources are limited.

4.3. Generalization capability

Consistent performance across in-domain (murine cerebrovascular) and out-of-domain (plant vein) test sets demonstrate Light-RepViTSR’s robust generalization. This capability, validated by statistically significant improvements (p < 0.05, Cohen’s d = 0.21–0.34) confirms clinical applicability across diverse biological structures and imaging conditions. The network’s ability to accurately reconstruct both fine capillary networks and complex fractal-like branching patterns suggests broad utility across various biomedical imaging applications.

4.4. Architectural insights

Ablation studies provide validated design principles for SR networks in medical imaging. The exclusion of SE layers, while counterintuitive for classification tasks, proves beneficial for super-resolution applications where channel-dependency modeling yields diminishing returns relative to computational cost. The identity mapping configuration provides sufficient representational capacity for capturing local and global dependencies essential to high-quality reconstruction while maximizing efficiency. These insights contribute to the broader understanding of efficient network design for medical image enhancement tasks.

4.5. Clinical implications

The combination of high-quality reconstruction and real-time processing capability addresses the fundamental limitation of PAE systems—the trade-off between spatial resolution and imaging speed. By enabling high-quality imaging from accelerated acquisitions, Light-RepViTSR facilitates high-speed, high-resolution in situ tumor biopsy. This could revolutionize clinical workflows by reducing diagnostic delays and enabling immediate intraprocedural pathological assessment, potentially improving patient outcomes through earlier and more accurate diagnosis.

4.6. Limitations and future directions

While Light-RepViTSR demonstrates significant advantages, several limitations warrant investigation. Performance variation between datasets suggests opportunities for domain adaptation techniques to further enhance cross-domain robustness. Future research should incorporate diverse pathological samples with clinical annotations to validate generalization in clinically relevant scenarios. Additionally, the exploration of adaptive scaling strategies for varying under-sampling patterns and the investigation of hardware-aware optimizations for specific endoscopic platforms will be valuable. Clinical validation in human subjects will be essential for translation to clinical practice, including the assessment of diagnostic accuracy and workflow integration.

5. Conclusion

We present Light-RepViTSR, an ultra-lightweight super-resolution network that adapts RepViT for real-time photoacoustic endoscopy. Rigorous evaluation on unseen test data demonstrates the following.

  • Superior reconstruction quality: significant improvements across scaling factors, with up to +1.41 dB PSNR enhancement at ×8 scaling versus conventional methods.

  • Unprecedented efficiency: 98.4%–99.0% model compression and 34%–56% inference acceleration compared to some methods.

  • Robust generalization: consistent performance across in-domain (cerebrovascular) and out-of-domain (plant vein) datasets.

Light-RepViTSR’s strategic integration of RepViT’s re-parameterizable convolutional blocks with task-specific optimizations establishes a new paradigm for efficient medical image enhancement. By eliminating non-essential components while preserving representational capacity, the network achieves an optimal balance between reconstruction quality and computational efficiency—critical for real-time clinical applications.

This study addresses the critical motion artifact bottleneck in PAE by enabling high-quality imaging from accelerated acquisitions. The demonstrated performance and efficiency advantages establish a clear pathway toward practical, real-time, high-resolution PAE for in situ tumor biopsy. Light-RepViTSR’s design principles extend to other resource-constrained medical imaging modalities, promising a broad impact in computational biomedicine and cutting-edge AI healthcare applications.

Future research will focus on clinical validation, hardware-specific optimization, and extension to multi-modal imaging, further advancing the frontier of real-time computational imaging in clinical oncology.

Funding Statement

The author(s) declared that financial support was received for this work and/or its publication. This work was supported in part by the General Program of the National Natural Science Foundation of China under Grant 62473192, in part by the Regional Cooperation Fund of Guangdong Province-Regional Cultivation Program under Grant 2024A151540132, and in part by Hong Kong RGC GRF under Grant 14220622.

Footnotes

Edited by: Enbo Zhu, University of California, Los Angeles, United States

Reviewed by: Boyi Li, Fudan University, China

Sudeep Mondal, University of South Florida, United States

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors without undue reservation.

Ethics statement

Ethical approval was not required for the study involving animals in accordance with the local legislation and institutional requirements because the dataset used for AI training is obtained from open source data from Duke PAM dataset (https://zenodo.org/records/4042171). Our experiment did not involve any research on animal samples.

Author contributions

GJ: Conceptualization, Funding acquisition, Methodology, Project administration, Writing – review and editing. RY: Data curation, Formal analysis, Investigation, Methodology, Software, Writing – original draft. ZF: Data curation, Formal analysis, Investigation, Methodology, Software, Writing – original draft. YL: Visualization, Writing – original draft. XY: Visualization, Writing – original draft. LL: Conceptualization, Funding acquisition, Methodology, Project administration, Resources, Writing – review and editing.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  1. Baad M., Lu Z. F., Reiser I., Paushter D. (2017). Clinical significance of us artifacts. Radiographics 37, 1408–1423. 10.1148/rg.2017160175 [DOI] [PubMed] [Google Scholar]
  2. Bashir S. M. A., Wang Y., Khan M., Niu Y. (2021). A comprehensive review of deep learning-based single image super-resolution. PeerJ Comput. Sci. 7, e621. 10.7717/peerj-cs.621 [DOI] [PMC free article] [PubMed] [Google Scholar] [Retracted]
  3. Chen M., Knox H. J., Tang Y., Liu W., Nie L., Chan J., et al. (2019). Simultaneous photoacoustic imaging of intravascular and tissue oxygenation. Opt. Lett. 44, 3773–3776. 10.1364/OL.44.003773 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Chen F., Li S., Han J., Ren F., Yang Z. (2024). Review of lightweight deep convolutional neural networks. Archives Comput. Methods Eng. 31, 1915–1937. 10.1007/s11831-023-10032-z [DOI] [Google Scholar]
  5. Ciepła J., Smolarczyk R. (2024). Tumor hypoxia unveiled: insights into microenvironment, detection tools and emerging therapies. Clin. Exp. Med. 24, 235. 10.1007/s10238-024-01501-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Cong S., Zhou Y. (2023). A review of convolutional neural network architectures and their optimizations. Artif. Intell. Rev. 56, 1905–1969. 10.1007/s10462-022-10213-5 [DOI] [Google Scholar]
  7. Das S., Dey M. K., Devireddy R., Gartia M. R. (2023). Biomarkers in cancer detection, diagnosis, and prognosis. Sensors 24, 37. 10.3390/s24010037 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Guo Y., Li F., Li K., Wang H., Xu P. (2025). Ssrepvm-unet: a lightweight hybrid model for medical image segmentation based on channel parallelism. Appl. Intell. 55, 911. 10.1007/s10489-025-06780-z [DOI] [Google Scholar]
  9. Haris M., Shakhnarovich G., Ukita N. (2018). Deep back-projection networks for super-resolution [DOI] [PubMed] [Google Scholar]
  10. Hu S., Wang L. V. (2010). Photoacoustic imaging and characterization of the microvasculature. J. Biomedical Optics 15, 011101. 10.1117/1.3281673 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Hussain S., Mubeen I., Ullah N., Shah S. S. U. D., Khan B. A., Zahoor M., et al. (2022). Modern diagnostic imaging technique applications and risk factors in the medical field: a review. BioMed Research International 2022, 5164970. 10.1155/2022/5164970 [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Kaur M., Lane P. M., Menon C. (2020). Endoscopic optical imaging technologies and devices for medical purposes: state of the art. Appl. Sci. 10, 6865. 10.3390/app10196865 [DOI] [Google Scholar]
  13. Krafft C., Popp J. (2023). Opportunities of optical and spectral technologies in intraoperative histopathology. Optica 10, 214–231. 10.1364/optica.478211 [DOI] [Google Scholar]
  14. Ledig C., Theis L., Huszar F., Caballero J., Cunningham A., Acosta A., et al. (2017). Photo-realistic single image super-resolution using a generative adversarial network [Google Scholar]
  15. Lim B., Son S., Kim H., Nah S., Lee K. M. (2017). Enhanced deep residual networks for single image super-resolution. [Google Scholar]
  16. Lin L., Wang L. V. (2022). The emerging role of photoacoustic imaging in clinical oncology. Nat. Rev. Clin. Oncol. 19, 365–384. 10.1038/s41571-022-00615-3 [DOI] [PubMed] [Google Scholar]
  17. Liu Y., Bhattarai P., Dai Z., Chen X. (2019). Photothermal therapy and photoacoustic imaging via nanotheranostics in fighting cancer. Chem. Soc. Rev. 48, 2053–2108. 10.1039/c8cs00618k [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Liu H.-I., Galindo M., Xie H., Wong L.-K., Shuai H.-H., Li Y.-H., et al. (2024). Lightweight deep learning for resource-constrained environments: a survey. ACM Comput. Surv. 56, 1–42. 10.1145/3657282 [DOI] [Google Scholar]
  19. Majidpoor J., Mortezaee K. (2021). Angiogenesis as a hallmark of solid tumors-clinical perspectives. Cell. Oncol. 44, 715–737. 10.1007/s13402-021-00602-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Mikami H., Gao L., Goda K. (2016). Ultrafast optical imaging technology: principles and applications of emerging methods. Nanophotonics 5, 497–509. 10.1515/nanoph-2016-0026 [DOI] [Google Scholar]
  21. Mirg S., Turner K. L., Chen H., Drew P. J., Kothapalli S.-R. (2022). Photoacoustic imaging for microcirculation. Microcirculation 29, e12776. 10.1111/micc.12776 [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Pain C. D., Egan G. F., Chen Z. (2022). Deep learning-based image reconstruction and post-processing methods in positron emission tomography for low-dose imaging and resolution enhancement. Eur. J. Nucl. Med. Mol. Imaging 49, 3098–3118. 10.1007/s00259-022-05746-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Park J., Choi S., Knieling F., Clingman B., Bohndiek S., Wang L. V., et al. (2025). Clinical translation of photoacoustic imaging. Nat. Rev. Bioeng. 3, 193–212. 10.1038/s44222-024-00240-y [DOI] [Google Scholar]
  24. Qiu D., Cheng Y., Wang X. (2023). Medical image super-resolution reconstruction algorithms based on deep learning: a survey. Comput. Methods Programs Biomed. 238, 107590. 10.1016/j.cmpb.2023.107590 [DOI] [PubMed] [Google Scholar]
  25. Rodrigues D., Barbosa A. I., Rebelo R., Kwon I. K., Reis R. L., Correlo V. M. (2020). Skin-integrated wearable systems and implantable biosensors: a comprehensive review. Biosensors 10, 79. 10.3390/bios10070079 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Roukos D. H., Murray S., Briasoulis E. (2007). Molecular genetic tools shape a roadmap towards a more accurate prognostic prediction and personalized management of cancer. Cancer Biology and Therapy 6, 308–312. 10.4161/cbt.6.3.3994 [DOI] [PubMed] [Google Scholar]
  27. Schiffman J. D., Fisher P. G., Gibbs P. (2015). Early detection of cancer: past, present, and future. Am. Soc. Clin. Oncol. Educ. Book 35, 57–65. 10.14694/EdBook_AM.2015.35.57 [DOI] [PubMed] [Google Scholar]
  28. Shin M., Seo M., Lee K., Yoon K. (2024). Super-resolution techniques for biomedical applications and challenges. Biomed. Eng. Lett. 14, 465–496. 10.1007/s13534-024-00365-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Song Q., Jin W., Qi J., Tan Y., Wen Z., Wang L., et al. (2025). Optical multimodal endoscopy. Laser and Photonics Rev. 19, e00286. 10.1002/lpor.202500286 [DOI] [Google Scholar]
  30. Wang A., Chen H., Lin Z., Han J., Ding G. (2023). Repvit-sam: towards real-time segmenting anything. arXiv Preprint arXiv:2312.05760. Available online at: https://api.semanticscholar.org/CorpusID:266162681. [Google Scholar]
  31. Wang A., Chen H., Lin Z., Han J., Ding G. (2024). “Repvit: revisiting mobile cnn from vit perspective,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 15909–15920. [Google Scholar]
  32. Wu J., Zhang K., Huang C., Ma Y., Ma R., Chen X., et al. (2024). Parallel diffusion models promote high detail-fidelity photoacoustic microscopy in sparse sampling. Opt. Express 32, 27574–27590. 10.1364/OE.528474 [DOI] [PubMed] [Google Scholar]
  33. Yang H., Wang Z., Liu X., Li C., Xin J., Wang Z. (2023). Deep learning in medical image super resolution: a review. Appl. Intell. 53, 20891–20916. 10.1007/s10489-023-04566-9 [DOI] [Google Scholar]
  34. Yoon T.-J., Cho Y.-S. (2013). Recent advances in photoacoustic endoscopy. World Journal Gastrointestinal Endoscopy 5, 534–539. 10.4253/wjge.v5.i11.534 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Zhang L., Li H., Liu X., Niu J., Wu J. (2020). “Mobilesr: efficient convolutional neural network for super-resolution,” in Globecom 2020 - 2020 IEEE global communications conference, 1–6. 10.1109/GLOBECOM42002.2020.9322623 [DOI] [Google Scholar]
  36. Zhang Y., Jin X., Zhang X., Wu Y., Tu L. (2025). Echomamba: a new mamba model for fast and efficient hyperspectral image classification. PloS One 20, e0330678. 10.1371/journal.pone.0330678 [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Zhao C., Shao M., Carass A., Li H., Dewey B. E., Ellingsen L. M., et al. (2019). Applications of a deep learning method for anti-aliasing and super-resolution in mri. Magn. Resonance Imaging 64, 132–141. 10.1016/j.mri.2019.05.038 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors without undue reservation.


Articles from Frontiers in Bioengineering and Biotechnology are provided here courtesy of Frontiers Media SA

RESOURCES