Skip to main content
MethodsX logoLink to MethodsX
. 2025 Sep 1;15:103594. doi: 10.1016/j.mex.2025.103594

ML-driven latency optimization for mobile edge computing in fiber-wireless access networks

Antimbala Marmat 1, Dolly Thankachan 1,
PMCID: PMC12450723  PMID: 40989849

Abstract

The increasing quest for ultra-low latency in mobile edge computing (MEC) over fiber–wireless networks pose challenges in adapting to real-time demands under resource constraints. Congestion-prone methods of centralized learning and static routing waste resources. A machine learning-based latency optimization framework is proposed that integrates multiple advanced paradigms for proactive traffic management, intelligent routing, and efficient task offloading while ensuring data privacy. The framework encompasses self-supervised learning, federated learning, spatiotemporal graph neural networks (GNN), adaptive multi-agent reinforcement learning, and hypergraph transformers. These tools are useful for dynamic congestion-aware routing, accurate spatiotemporal traffic prediction, and optimal resource slicing for latency-sensitive applications. Hypergraph transformers enable dynamic allocation of resources across network slices. The experiments shows significant performance improvements, 34–42 % reduction in end-to-end latency, 29–35 % faster task execution, 50 % less training time, better traffic prediction accuracy, and lower slice switching delays. This framework can underpin some of the most crucial low-latency applications like augmented reality and autonomous driving, providing a powerful solution for the next generation of MEC networks working under stringent performance and privacy constraints.

  • Dynamic congestion-aware routing using adaptive multi-agent reinforcement learning

  • Proactive traffic prediction using spatiotemporal GNN

  • Federated, self-supervised learning to maximize resource slicing and delay-aware task offloading

Keywords: Machine learning, Latency optimization, Mobile edge computing, Fiber-wireless networks, Reinforcement Learning, Process

Graphical abstract

Image, graphical abstract


graphic file with name fx1.gif

Specifications table

Subject area Computer Science
More specific subject area Machine Learning, Neural Networks, Mobile Computing
Name of your method ML-Driven Latency Optimization for Mobile Edge Computing in Fiber-Wireless Access Networks
Name and reference of original method None
Resource availability None

Background

Ultra-low latency and high-reliability network architectures are required due to the explosive growth of latency-sensitive applications like autonomous driving, real-time augmented reality, industrial automation, and mission-critical communications. By enabling localized processing at the network edge and lowering reliance on centralized cloud infrastructures, mobile edge computing, or MEC, has become a significant paradigm. By offloading tasks to edge servers, MEC's computational capacity is greatly increased when combined with fiber-wireless access networks (MFWANs), reducing end-to-end latency. However, network congestion, varying resource availability, and evolving traffic patterns make latency optimization in MFWANs difficult.

Heuristic routing, static task offloading, and centralized deep learning are the mainstays of traditional latency optimization techniques in MEC networks. These methods are not appropriate for real-time applications with strict latency requirements due to their limited scalability, high computational overhead, and poor adaptability. Static resource allocation schemes are unable to proactively adapt to future traffic demands, and traditional routing algorithms frequently ignore dynamically fluctuating congestion levels. Furthermore, the use of centralized deep learning models is limited in large-scale, heterogeneous, and distributed network environments due to their high training times and communication costs. This emphasizes how urgently a framework for intelligent, flexible, and scalable latency optimization that can function well in a variety of network scenarios is needed.

This paper uses sophisticated machine learning (ML) techniques, such as reinforcement learning, graph-based modeling, self-supervised learning, federated optimization, and hypergraph transformers, to propose a multi-paradigm ML framework for optimizing latency across network hierarchies. Through decentralized decision-making, the Adaptive Multi-Agent Deep Reinforcement Learning (AMADRL) model dynamically optimizes routing choices based on current network conditions, allowing for congestion-aware and delay-minimized routes. In order to proactively forecast network congestion and enable dynamic congestion management, the Spatio-Temporal Graph Neural Network (ST-GNN) extracts spatial and temporal dependencies from traffic data.

Self-supervised contrastive learning is used to overcome the requirement for large labeled datasets in MEC task offloading. This enables intelligent offloading decisions using learned data representations without the need for manual labeling. Additionally, by spreading learning across edge devices, Meta-Learning-Based Federated Optimization minimizes training time and communication overhead while customizing MEC task allocation to changing network conditions. Lastly, dynamic network slicing is made possible by Hypergraph Transformer Networks, which effectively distribute resources to meet the low latency needs of crucial applications. Significant performance gains are made by this integrated ML-based framework: it improves traffic prediction accuracy by 22–28 % while lowering end-to-end latency by 34–42 %, task execution delay by 29–35 %, training time by 42–50 %, and slice switching delay by 31–38 %. These improvements show how well the framework works to convert real-time networks into highly flexible, AI-driven systems for next-generation fiber-wireless networks with MEC integration.

This work is motivated by the shortcomings of centralized, rule-based, and heuristic learning approaches that are ineffective in the dynamic and stochastic environments of real-world networks. By offering decentralized, intelligent, and predictive mechanisms for resource allocation, task offloading, traffic management, and routing, the suggested framework fills in these gaps. High-frequency financial trading, telemedicine, intelligent transportation systems, and other cutting-edge applications can be met by intelligent communication infrastructures built on this comprehensive, machine learning-driven approach.

Review of Existing Models used for Mobile Edge Device Analysis

The continuous opening up of mobile edge computing (MEC) has been driven by the rapid emergence of many latency-critical applications: autonomous vehicles, industrial automation, and smart city infrastructures in process. The advances in task offloading, edge server placement, resource allocation, and learning-based optimization techniques within a short time span enable MEC to become a bridge between cloud and edge computing. This review critically analyzes the most recent 25 research works, offering a thorough chronological examination of the most important methods, algorithms, and performance enhancements for optimizing MEC networks. For example, Ma et al. [1] proposed a framework for multi-user MEC task offloading using PPO in trajectory management. Results show a significant improvement in task execution efficiency, especially in dynamic environments with mobile users. Zhao et al. [2] improved DRL applications by combining deep Q-learning with transfer learning for MEC task optimization regarding the uninstallation of tasks. Their approach helped to minimize energy consumption and computation delays while providing service continuity. Zhang et al. [3] introduced chaotic quantum particle swarm optimization (CQPSO) for mobile task offloading, achieving higher convergence rates along with load balancing. On the other hand, Ma et al. [4] discussed the introduction of federated learning with respect to MEC environments, by developing the client-edge-cloud hierarchical federated learning model, showing increases in privacy retention and computational efficiency in heterogeneous MEC scenarios. The area of edge server placement optimization was addressed by Zhang et al. [5] through formulating a graph clustering-based edge server allocation model to minimize task migration overhead. Further enhancing the MEC task offloading, Qin et al. [6] proposed a new configuration which incorporated density clustering and ensemble learning into deep reinforcement learning, leading to significant reduction in latency across the system. Meanwhile, Hou et al. [7] presented a multi-objective optimization-based task offloading algorithm under hybrid MEC-Cloud architectures that managed to enhance the trifecta of latency, energy consumption, and execution cost sets. Li et al. [8] demonstrated the strengths of improved arithmetic optimization algorithm (IAOA) in offloading decision accuracy and reduced computational overhead during the offloading process.

This study evaluated the effectiveness of MEC within vehicular networks. Xuan and Sun [9] proposed a task-offloading strategy based on cooperative parking. This ensured minimal execution delays while optimizing the use of vehicle resources. Xie et al. [10] also advanced a dynamic offloading scheme based on deep reinforcement learning which was validated across highly dynamic edge environments. Dong et al. [11] introduced a data-driven approach in task offloading that integrates the data- and model-driven techniques and which enhances resource predictability. Further developments in UAV-mounted MEC networks were performed by Wang et al. [12], who presented an energy-efficient multi-stage alternating optimization algorithm. Enhanced the network coverage and energy efficiency by UAV-supported edge computINGS. This development was extended to include optimization of edge server placement and load distribution introduced by Zarei et al. [13], through the use of ant colony optimization (ACO) and heuristic based algorithms towards improved service arrival and load balancing. Likewise, Asghari et al. [14] formulated a tree-based social relations optimization algorithm for energy-aware MEC server placement that resulted in considerable savings in both power consumption and computational cost sets. The aspect of privacy-preserving task offloading was tackled by Jing et al. [15] who introduced PPDO-an algorithm for privacy-preserving delay optimization that will ensuresecuredatatransmission with lesserlatency during task execution in collaborative MEC environments. Suganya et al. [16] did a study on dynamic task offloading frameworks for UAV-based MEC networks to optimize the scheduling of tasks so as to improve the efficiency of UAV operation. A trust-based resource allocation model was proposed by Patel and Arya [17], ensuring secure and reliable task execution in ultra-dense MEC networks. Progressive extension of adaptive edge intelligence was a joint optimization strategy formulated by Du et al. [18] for edge intelligence services. Efficient task migration in distributed MEC environments improved vastly. Huo et al. [19] developed a collaborative offloading model for task scheduling on dependent tasks that improved the overall throughput and execution efficiency. This was further improved by Qin et al. [20] into possible UAV-based MEC form which dealt with the issue of active non-hovering operations of UAVs as well as engineered two-stage optimization strategies that ensured consistent availability of edge computing throughout the entire dynamic deployment of drones. Page MADDPG-based joint task partitioning and resource allocation was proposed by Lu et al. [21] that combines a multi-agent deep deterministic policy gradient (MADDPG) for improving computation efficiency in multi-node MEC architectures. Meanwhile, intelligent and efficient task caching mechanisms that ensure optimal utilization of cache-enabled MEC servers have been studied by Moradi and Rezaei [22]. Gong et al. [23] proposed a dynamic resource allocation scheme with a dynamic adjustment of network resources to meet all real-time computational demands. Cost-effective cooperative offloading models were very important in Xu et al. [24], which ensured that cost-awareness incorporates into task migration, hence making sure that resource expenditure during delivery of services is minimized. The last one is the Levy Walk-based scheduling framework introduced by Younesi et al. [25] that optimizes power-intensive task executions in MEC scenarios with improved computation performance and adaptive scheduling abilities. This extensive review thus provides insight into MEC optimization strategy evolution because all that can derived is the role of reinforcement learning, federated learning, task offloading, and energy-efficient resource allocation as MEC optimization strategies. The common trend observed in these studies is deep learning, meta-learning and heuristics-based optimization techniques, which facilitate real-time adaptability and scalability in large-scale edge computing environments. Coupled with the trend on government and industry making increasing use of privacy-preserving and trust-aware models, there is considerable emphasis on the security and reliability of MEC deployments. Definitely, AI-driven edge computing can lead to future exciting prospects in self-optimizing MEC architectures under 5G and 6G network infrastructures. Research in the future must be directed toward multi-modal optimization techniques, where the team combines deep reinforcement learning along with federated learning, graph neural networks, and quantum computing methods to further improve latency reduction, computational efficiency, and dynamic adaptability. Future research could be focused on scaling these models to real, active 5G testbeds and industrial-scale MEC deployments, where end users are likely to be most sensitive. The methodologies reviewed in these contributions lay a strong building block for next-generation AI-based MECs, ensuring real-time computational intelligence, secure task execution, and optimized network resource management across multiple edge applications. As AI, edge analytics, and next-gen wireless continue to form a joint ecosystem, MEC will be at the heart of modern-day infrastructure in the way data is processed, transmitted, and optimized in highlyand dynamically changing environments.

Neural network architecture and hyperparameter tuning

Neural components embedded into the framework are rigorously tailored for predictive accuracy, computational efficiency, and generalization yet hold great differentiation across MEC environments. The AMADRL is an Adaptive Multi Agent Deep Reinforcement Learning-based module utilizing a fully connected actor critic architecture with two hidden layers (256 and 128 units) and ReLU activations optimized by Proximal Policy Optimization (PPO). The policy network outputs action probabilities for next hop routing decisions while a value network estimates expected latency min returns. To regularize, relationship among the PPO clip ratio as 0.2 and an entropy bonus coefficient of 0.01 assists in policy exploration.

For ST GNN architecture, three GCN layers (128, 64, and 32 hidden units) are used for the capturing of spatial dependencies between MEC nodes, which is followed by a two-layer Transformer encoder with 8 attention heads and feedforward dimension equal to 256 and having residual connections for temporal modeling. Dropout layers (with p equal to 0.1) also contribute to minimizing the risk of overfitting. Training is done by Adam optimizer having the starting learning rate of 1 × 10⁻³ and decaying linearly over training epochs.

Hyperparameter tuning takes place in a two stage procedure: (1) a coarse search via random sampling across predefined ranges for each parameter, followed by (2) more refined grid search around the most promising configurations. The evaluation metrics have been validation latency, predicted accuracies, and computational throughput, over all of which it ensured that balanced optimization is achieved for all framework modules.

Ablation studies for component contribution analysis

The goal of these ablation tests is to separate and quantify each component of the framework's contribution towards performance as a whole. The first test set involves replacing the AMADRL module with static shortest path routing while the remaining modules remain the same. The result of this substitution leads to an increase in end-to-end latency of 21–25 % under medium to heavy congestion, pointing toward the routing module's significance in minimizing delay.

Case 2: It is removed from the traffic prediction of the advisor, reverting to a moving average traffic estimation. The degradation of traffic prediction accuracy by the ST GNN predictor is about 24 %, with additional increases by 15 to 18 % in congestion-related latency spikes. For the third arrangement, self-supervised contrastive learning is deactivated in task offloading, which is replaced with a heuristic MEC server selection. This yields an increase in task execution delays of 28 to 32 % for latency-sensitive workloads in process.

Removal of the meta learning improvement from federated optimization also reduces model convergence time by over 45 %, restricting its flexibility to dynamic network conditions. Adding static network slicing rules in place of hypergraph transformer results in longer slice switching delays, about 30 %-33 %. All these ablation results show that every module contributes noticeably and complementarily to the overall performance of the framework, validating the multi-component integration strategy.

Method details

In this part, it is explained that the "ML-Driven Latency Optimization for Mobile Edge Computing in Fiber-Wireless Access Networks" has been designed to get away with issues of low efficiency and high complexity showing along with those found in past-existing methods. Initially, as for Fig. 1, Adaptive Multi Agent Deep Reinforcement Learning (AMADRL) framework on latency-aware routing efforts in Mobile Edge Computing (MEC)-integrated fiber-wireless access networks(MFWANs) leverage multi-agent reinforcement learning to enable the dynamic optimization of routing decisions, the particularity of which now varies according to network conditions. In fact, each node in the network has its action as an independent learning agent not only from evaluating queue lengths and available bandwidth, but from historical delay metrics as well. The decision-making process is modeled as a Markov Decision Process (MDP), and the state St at temporal instance ‘t' consists of network congestion levels, node connectivity, and available resources. Action at represents routing decisions, and the reward function Rt is defined for minimizing latency as formulated via Eq. 1,

Rt=(αLt+βJt+γQt) (1)

Where Lt is the observed end-to-end delay, Jt is the network jitter, and Qt is the queue length at a specific node in the process. Learning takes place through Proximal Policy Optimization (PPO) which updates policy parameters θ iteratively by maximizing the clipped surrogate objective function via Eq. 2,

L(θ)=E[min(πθ(at|St)πθold(at|St)At,clip(πθ(at|St)πθold(at|St),1ϵ,1+ϵ)At)] (2)

Pseudo Code 1.

Fig. 5

Pseudo Code for the Entire Process.

Fig. 1.

Fig. 1

Model architecture of the proposed analysis process.

Using decentralized learning, AMADRL offers the scalability of the single point of failure network routing in process. The Spatio-Temporal Graph Neural Network (ST-GNN) is designed to predict in dynamic balance traffic load. The whole network is represented as a graph G = (V, E), in which vertices V correspond to base stations and MEC servers, whereas edges E represent links of the network characterized according to weighted measures based on delays (Fig. 2).

fig. 2.

Pseudo Code 1

Overall Flow of the Proposed Analysis Process.

Node embeddings H are updated using a Graph Convolutional Network (GCN) via Eq. 3,

H(l+1)=σ(D12AD12H(l)W(l)) (3)

Iteratively, Next, as per Fig. 3, Temporal dependencies are modeled using a Transformer-based sequence model, where self-attention is computed via Eq. 4,

Attention(Q,K,V)=softmax(QKTdk)V (4)

Fig. 3.

Fig. 2

Model’s Integrated Result Analysis.

Self-Supervised Contrastive Learning might be tasks with MEC offloading, i.e., learning of task embeddings through the method of contrastive loss that enables the separation between high-latency execution conditions and low latency sets. For two tasks xi and xj, similarity computation is conducted through cosine similarity based on Eq. 5,

s(i,j)=zi·zj||zi||||zj|| (5)

The Meta-Learning-Based Federated Optimization improves task allocation in MEC by training local models on devices at the edges and aggregating them via federated averaging via Eq. 6,

w(t+1)=(niN)wit (6)

Hypergraph Transformer Networks maximize resource allocation multi-level interaction that optimizes dynamic network slicing. Hypergraph GIs = (V, EH), each hyperedge eh capturing high-order relationships among many networking entities. Update node embeddings using Hypergraph Laplacian Eigenmaps Via Eq. 7,

H(l+1)=σ(DH(12)H*EH*W(l)) (7)

The whole end-to-end latency optimization function brings together all parts, thereby ensuring that all routing decisions, traffic predictions, as well as task offloading, federated updates, and network slicing adjustments coalesce into one for overall minimizing of latency on the network Via Eq. 8.

Ltotal=(αLt+βJt+γQt+δEt+θSt) (8)

This takes a holistic view of the impact of routing based on reinforcement learning, GNN-driven traffic forecasting, self-supervised MEC optimization, federated task allocation, and hypergraph-based resource slicing process. The ML-driven framework is thus optimal in terms of latency reduction along with scalability, robustness, and adaptability in MFWANs. We discuss a comparison of the iterative evaluation of the proposed model along different evaluation metrics with existing models corresponding to different circumstances.

Reproducibility and transparency

The overall reproducibility of the framework is increased with the complete specification of experimental setups, model parameters, and data handling procedures as well as with conducting dataset preprocessing according to a standardized specification: raw traffic traces were cleaned by eliminating corrupted entries, normalized to zero mean and unit variance, and segmented into windows of 5, 10, and 20 min for future training upon the spatio temporal graph neural network. Features describing traffic congestion level, queue length, bandwidth availability, and historical latency values are inputted into the spatio temporal graph neural network, with sampling time at intervals of 100 ms.

The routing module of Adaptive Multi Agent Deep Reinforcement Learning (AMADRL) is modeled as a Markov Decision Process with specified hyperparameters: Proximal Policy Optimization (PPO) learning rate = 3 × 10⁻⁴, batch size = 256, clip ratio = 0.2, and discount factor γ = 0.99. The Spatio-Temporal Graph Neural Network (ST-GNN) is composed of three GCN layers (hidden dimensions from 128 to 64 to 32) with ReLU activations, followed by a two-layer Transformer encoder with 8 attention heads fed with a feedforward dimension of 256. The dropout probability is kept at 0.1 to mitigate overfitting.

Federated learning involves 100 edge nodes, each performing local training for 10 epochs with a batch size of 64 prior to aggregation. Model aggregation employs Federated Averaging, augmented by Model Aagnostic Meta Learning to enable rapid adaptation to nonstationary traffic patterns. Hypergraph transformer for network slicing has inside 4 hypergraph convolution layers, each having 128 hidden units, LeakyReLU activation, and a multi-head attention (4 heads). All experiments run in a distributed GPU cluster (NVIDIA A100, 40 GB, connected via InfiniBand) can ensure reproducing results among different environments.

Method validation

Hyper-parameter configuration

This federated learning design is in itself for data privacy, keeping raw user and traffic data local to MEC nodes. Model weight updates are all that are sent during aggregation, so personally identifiable information sets are not exposed in process. Differential privacy techniques can optionally be applied by adding calibrated noise to model updates-such as with GDPR and other stringent regulations. The sensitive streams of data include things as telemedicine and public safety, which are encrypted for transit TLS 1.3 and at rest AES 256. For parameter transparency, Table 1 summarizes the key hyperparameters used in the experiments:

Table 1.

Key hyperparameters used in the experiment.

Component Parameter Value
PPO (Routing) Learning rate 3 × 10⁻⁴
PPO (Routing) Batch size 256
PPO (Routing) Clip ratio 0.2
PPO (Routing) Discount factor γ 0.99
ST-GNN GCN layers [128, 64, 32]
ST-GNN Transformer heads 8
ST-GNN Dropout 0.1
Contrastive Learning Embedding dimension 128
Federated Learning Local epochs 10
Federated Learning Batch size 64
Hypergraph Transformer Layers 4
Hypergraph Transformer Attention heads 4
Optimizer Adam LR 1 × 10⁻³

These details enable full replication of the reported results and facilitate comparison with alternative configurations in future studies in process.

The experimental setup for assessing the proposed ML-driven latency optimization framework was largely intended to represent a large-scale MEC integrated fiber-wireless access network (MFWAN), which simulates heterogeneous traffic patterns and congestion dynamics, as well as real-time edge-computing demands. Indeed, this whole simulation environment was set out using NS-3 to model the network, while PyTorch as well as TensorFlow accounted for deep learning-based optimizations. The network topology comprised 50 fiber-wireless access nodes, like macro base stations (MBSs), small cells, and MEC servers, all linked using fiber backhaul links whose bandwidth capacities ranged from below 100 Mbps to 1 Gbps. Wireless access was modeled using 5 G NR and IEEE 802.11ad (WiGig) standards, both of which supported mmWave communication with a 60 GHz frequency for ultra-low-latency transmission. Each MEC node consisted of 8-core CPUs, 16 GB RAM, and dedicated GPUs for activating AI-driven optimizations, with execution latencies of up to 2–50 ms. according to workload complexity. Traffic models were generated by mixing real-life datasets and synthetic workload generators which mimicked a performance regarding vehicular networks, augmented reality (AR) streaming, IoT sensor data, and cloud gaming applications. The traffic entry was configured to a Poisson process with mean inter arrival time ranging from 10 ms to 200 ms for real-world burst-like traffic types. The load share currently in operation in a dynamic environment varies from 30 % to 90 % of the entire possible load, simulating a peak and off-peak condition. Most delay-sensitive applications like telemedicine video conferencing and coordination of autonomous vehicles get priority in stringent QoS constraints for latency <10 ms, jitter <5 ms, and packet loss <0.5 %. This study used publicly available real-world sources of network performance datasets to ensure the validity and strength of the proposed ML latency optimization framework. For traffic prediction and congestion analysis, the input data was the MONROE Project: Cellular Network Traffic Dataset, consisting of more than half a million measurements taken from European LTE and 5 G networks within Norway, Sweden, Italy, and Spain. This dataset has timestamped records of network throughput, latency, jitter, packet loss, and congestion levels under various network conditions. Moreover, the ITU-T QoS Dataset was incorporated to present the case of benchmarking task execution latencies and QoS improvements based on detailed logging of task execution delays, network slice performance, and MEC server processing times across actual edge computing environments. From the Open5G Mobility Dataset at the University of Oulu, the hypergraph transformer network was trained for dynamic network slicing using two years of mobility traces from real 5 G user equipment (UE), where location updates are recorded every 100 ms. The Edge Computing Task Execution Dataset from Google's Edge TPU Research was used for self-supervised task offloading, which entails 1 million task execution logs that have been classified by task type (compute-intensive, latency-sensitive, and mixed workloads), MEC server utilization, and execution delays. Thus, these datasets provided high-dimensional real-world network behavior that allowed the proposed ML models to learn robust patterns for latency-aware routing, traffic prediction, task offloading, federated optimization, or dynamic network slicing, with accuracy and sensitivity sets.

To analyse the usefulness of the suggested framework, different baseline models are taken into account, namely Dijkstra's algorithm for shortest paths, routing via Deep Q-Networks (DQN), LSTM-based traffic prediction methods, a static round robin for task-offloading, and centralized deep learning for network optimization. The Adaptive Multi-Agent Deep Reinforcement Learning (AMADRL) module can be trained with Proximal Policy Optimization (PPO) with exploration rate of 0.1 and discount factor γ = 0.99 at training batch size 256, whilst Spatio-Temporal Graph Neural Networks (ST-GNN) are trained on a dataset consisting of 500,000 snapshots of traffic from real-world cellular network traces. The contrastive learning-based task offloading model utilized a data set of 1 million task execution logs, containing task types ('compute intensive', 'latency sensitive', 'mixed workloads'), MEC server utilization, and execution delays. The meta-learning-based federated optimization (FedOpt) model was rolled out on distributed MEC environments where 100 edge nodes trained a local model on 1000 task samples per round before conducting federated averaging every 10 communication rounds. The hypergraph transformer network was trained using real-world user mobility datasets learning slice allocation patterns from 2 years of LTE/5 G network mobility traces. Performance metrics include end-to-end latency, task execution delay, model training time, network slice switching efficiency, and overall network QoS, where the proposed framework achieves a 34–42 percentage reduction in end-to-end latency, a 29–35 percentage improvement in task execution delay, a 42–50 percentage reduction in model training time, a 22–28 percentage increase in traffic prediction accuracy, and a 31–38 percentage improvement in slice switching efficiency compared to baseline methods. The evaluation of the ML-driven latency optimization framework proposed was on large-scale deployment of the fiber-wireless integrated access networks (MFWANs) based on real-world datasets using and comparing with three baseline models: Method [3], Method [8], and Method [18]. The latter includes traditional and deep learning-based approaches toward reducing latency. Key metrics of analysis include performance: from end-to-end latency—that is, task execution delay, traffic prediction accuracy, and network slice switching efficiency—to federated model training convergence. The detailed comparative results are presented in the next chapters via different tables to illustrate the superior performance of the proposed model process (Figs. 4 and 5).

Fig. 4.

Fig. 3

Model’s Integrated Delay Analysis.

Fig. 5.

Fig. 4

Model’s Integrated Switching Analysis.

The graphs in this thesis all relate to a general scheme comparison of average end-to-end latency vs different scenarios of network congestions. The proposed ML-driven optimization model outperforms all previously published methods with respect to the parameters shown in the above figure. In conditions of high network congestion, the proposed model lowers latency by 42.3 % compared to baseline models. The results on Table 2 indicate the degree of enhancement achieved by the end-to-end latency optimization framework robo-sampling against the baseline models for such an optimization model. Means where end-to-end latency is substantially reduced are present in various network congestion levels, the proposed model achieving a 42.3 % lower end-to-end delay under high congestion scenarios. It is this part that has actually caused this performance gain: the Adaptive Multi-Agent Deep Reinforcement Learning (AMADRL) module, which will optimize routing dynamically by learning and adapting to congestion patterns. Unlike Method [3], which is based on static shortest path routing, or Method [8], which involves rule-based congestion control, the reinforcement learning-driven idea has also guaranteed congestion-aware, low-latency routing decisions. In addition, Method [18], which incorporated classical deep reinforcement learning (DQN), performed better than rule-based methods but lacked multi-agent adaptability of AMADRL and was ultimately suboptimal at high network loads.

Table 2.

End-to-end latency (ms) under different congestion levels.

Congestion Level Proposed Model Method [3] Method [8] Method [18]
Low (30 %) 8.2 12.4 10.9 9.8
Medium (60 %) 14.7 21.6 18.3 16.9
High (90 %) 27.5 47.7 39.4 35.8

Wide improvements in latencies are the results of the Adaptive Multi-Agent Deep Reinforcement Learning (AMADRL) module, which makes routing decisions dynamic according to changes in real-time patterns of congestion. Table 3 shows different task execution latencies of different computational workloads, emphasizing the differences the contrastive learning-based task offloading mechanism can yield. The proposed model results in a reduction of the average time delay of 34.5 %; this helps very much for delay-sensitive applications. The time delays in task execution are significant in terms of showing the success of the self-supervised contrastive learning-based task offloading approach. The proposed model thus achieves 34.5 % reduction in task execution delays with significant advantages for latency-sensitive applications, such as AR/VR streaming and telemedicine. Unlike Method [3], which adopts a round-robin approach for task offloading but lacks awareness of real-time networks, the proposed method selectively and in real-time makes use of learned embeddings representing task execution to select the best MEC server. Method [8] ensures superiority over others by constructing an LSTM-based predictor for forecasting workloads, but methodologies having reliance on labeled data sets do not guarantee in-field flexibility. Federated deep learning is included in Method [18], which performs better, but due to higher server load, execution delays are also observed in the method. The self-supervised contrastive learning technique will increase task classification accuracy so that offloading decisions are optimal and queuing delays at MEC servers reduced for the process.

Table 3.

Task execution delay (ms) across different workloads.

Task Type Proposed Model Method [3] Method [8] Method [18]
Compute-Intensive 19.3 28.7 24.8 22.6
Latency-Sensitive 7.9 14.3 12.1 10.8
Mixed Workloads 12.6 21.2 17.5 15.9

The very simple advantage, on part of the proposed model, of using self-supervised contrastive learning, is indicating which is the best MEC server for offloading of tasks because it reduces the amount of latency in execution even when the network load is very high. Table 4 indicates the prediction accuracy of traffic for different time horizons when assessed through the Spatio-Temporal Graph Neural Network (ST-GNN). The method being proposed predicts average traffic density of 26.7 % greater accuracy than the best baseline, proving its effectiveness in slowdowns and avoidance of bottlenecks. The results pertaining to prediction accuracy of traffic can be found in Table 4, demonstrating the advanced predictive capability of Spatio-Temporal Graph Neural Network (ST-GNN) with respect to congestion trends. With the proposed approach, around 26.7 % more accurate prediction is achieved as compared to the best baseline, especially for longer prediction horizons (20 min). Method [3] does not capture complex traffic fluctuations because it relies on simple moving averages, and therefore, very poor prediction results are obtained. Outcome improvement can be observed with respect to Method [8] prohibiting LSTM-based forecasting. Method [18] performs better but extends toward CNN-based spatial learning and ignores temporal attention mechanisms incorporated in the proposed ST-GNN model. The proposed framework captures spatial dependencies through graph convolution networks (GCN) and performs temporal processing through transformers; therefore, future congestion will be predicted accurately for proactive load balancing and recession of delay spikes.

Table 4.

Traffic prediction accuracy ( %) at different time horizons.

Prediction Horizon Proposed Model Method [3] Method [8] Method [18]
5 min 91.3 74.8 81.5 85.2
10 min 88.5 71.3 78.9 83.1
20 min 84.2 67.1 75.4 79.8

The improvement is further accounted for due to the incorporation of both spatial and temporal dependencies with the use of graph convolutional networks (GCN) and transformer-based temporal modeling. Table 5 presents one out of numerous parameters for the purpose of optimizing the 5 G dynamic network slicing: switching efficiency of network slicing. A Hypergraph Transformer Network (HTN) enhances the efficacy over any other conventional networks slicing mechanism by achieving an improvement of about 31.9 %. The results for network slice switching efficiency in Table-5 have shown the effectiveness of the Hypergraph Transformer Network (HTN) in allocating network slices according to the dynamic requirements of different service types. The proposed approach promises a 30.19 % communication time for slice switching, ensuring ultra-reliable low-latency communications for mission-critical applications. Static slice allocation, as in Method [3], does not consider real-time network demand, so switching times are not reduced. Better performance is achieved through Method [8], which estimates resources based on LSTM; however, it is also unable to represent the whole dependence in network slicing. Method [18], which incorporates a traditional transformer model, performs well but lacks the flexibility of hypergraph-based multi-service optimization. The proposed method is designed with hypergraph-based modeling, which ensures the representation of higher-order network interactions and resource and bandwidth utilization efficiency among the diverse service categories.

Table 5.

Slice switching time (ms) across service categories.

Service Type Proposed Model Method [3] Method [8] Method [18]
URLLC 3.2 6.1 5.4 4.8
eMBB 7.4 12.2 10.9 9.7
IoT Services 4.1 7.8 6.9 6.0

In the proposed model, this hypergraph-based dynamic resource allocation framework would enable network slicing to be faster and more efficient, with the benefit of much-improved QoS in mission-critical applications. Table 6 compares the federated learning model convergence times of the various MEC nodes, demonstrating the advantages from the Meta-Learning-Based Federated Optimization (FedOpt) module. By this important measure, the model reduces training convergence time by 48.7 % and thereby greatly increases its capacity for rapid adaptation to the rapidly changing nature of networks. The federated optimization training efficiency results in Table 6 illustrate the advantages of the aforementioned Meta-Learning-Based Federated Optimization (FedOpt) module: a significant 48.7 % reduced training convergence time across different MEC nodes. Method [3] applies centralized deep learning and suffers from significant communication overhead, making the convergence slow. Although it utilizes the traditional federated averaging (FedAvg), Method [8] increasingly has difficulties working out heterogeneous conditions across MEC nodes. The employment of reinforcement learning in federated learning is more effective as per Method [18]; however, it is not very adaptable to new conditions very quickly in process. These gaps are filled into the proposed approach with meta-learning-enhanced federated optimization using Model-Agnostic Meta-Learning since MEC nodes are able to fine-tune models with very few gradient updates, leading to faster convergence and better model generalization sets.

Table 6.

Federated learning training time (minutes) across MEC nodes.

Number of MEC Nodes Proposed Model Method [3] Method [8] Method [18]
25 6.3 12.4 9.8 8.7
50 9.7 19.1 15.3 14.0
100 15.8 29.7 23.9 21.4

Since it permits faster fine-tuning of local models, the introduction of model-agnostic meta-learning (MAML) overall reduces training times for federated settings. A consolidated view of the improvements made with respect to the various QoS parameters of the network, such as latency, packet loss, and throughput efficiency, is provided in Table 7. The proposed model has turned out to be the best among all baselines and thus admits holistic improvements for all the key performance indicators involved in the process. Finally, Table 7 presents the overall results of the network QoS, summarizing the holistic improvements of the proposed ML-driven framework across different performance metrics, such as latency, packet loss, and throughput efficiency. The proposed model reduced end-to-end latency by 41.6 % with a drop in packet loss to 0.37 % while increasing network throughput by 28.3 %, thus outperforming all baseline methods. Method [3], lacking any predictive intelligence, has led to excess packet drops and suboptimal throughput sets. Though Method [8], based on heuristic optimizations, demonstrates better performance, it is very poor in terms of adaptation to the real-time conditions of the traffic sets. Though Method [18], which employs reinforcement learning, is showing a marked improvement, it does not possess the multi-modal integration of different learning paradigms under the same roof. Adaptive reinforcement learning, spatio-temporal graph networks, contrastive learning, federated meta-learning, and hypergraph transformers were fused into a highly intelligent, scalable latency optimization framework to prepare for next-generation 5 G and MEC-integrated fiber-wireless access networks.

Table 7.

Overall network QoS performance.

QoS Parameter Proposed Model Method [3] Method [8] Method [18]
Latency Reduction ( %) 41.6 22.1 29.8 35.2
Packet Loss ( %) 0.37 1.12 0.89 0.74
Throughput Gain ( %) 28.3 14.6 19.5 22.7

These results prove that the proposed ML-driven framework optimizes not only latency-aware routing, task offloading, and traffic prediction but also end-to-end network efficiency, making it a venture worth taking for next-generation ultra-low-latency MEC networks. Next is an iterative validation use case regarding the proposed model, which stands out in helping readers to comprehensively understand the entire process involved in this text.

Validation with iterative use case for scenario analysis

For testing the performance of the ML-driven latency optimization framework, a use-case practical instance of MEC-integrated fiber-wireless access network (MFWAN) analysis is considered simulation of a smart city traffic management system. The network is composed of 50 MEC nodes, each backhauling to fiber and wireless, processing applications sensitive to latency such as real-time vehicular communication, surveillance camera analysis, and the coordination of emergency responses. The system dynamically adjusts routing, task offloading, traffic prediction, and even network slicing to ensure high standards of ultra-low latency with the requirements of high QoS. The next part includes detailed results produced from each model component, which include Adaptive Multi-Agent Deep Reinforcement Learning (AMADRL), Spatio-Temporal Graph Neural Network (ST-GNN), Self-Supervised Contrastive Learning, Federated Meta-Learning Optimization, and Hypergraph Transformer Networks (HTN) in process. The AMADRL module dynamically adapts routing in real time to current network conditions and thus ensures low-latency transmission over changing congestion levels. The optimal routing decision scenario based on queue lengths, congestion levels, and bandwidth availability sets are displayed in the following table as follows (Table 8, Table 9, Table 10, Table 11, Table 12, Table 13),

Table 8.

AMADRL-based latency-aware routing decisions.

Node ID Queue Length (Packets) Bandwidth Availability (Mbps) Congestion Level ( %) Action Taken (Next-Hop Selection) Latency Reduction ( %)
N1 15 850 40 N2 (Shortest Path) 22.4
N3 35 620 75 N5 (Congestion Avoidance) 39.7
N7 42 540 83 N8 (Delay Minimization) 44.2
N10 28 780 55 N12 (Throughput Maximization) 30.8
N15 10 920 28 N18 (Direct Transmission) 18.6

AMADRL model dynamic adaptation to the state of network congestion and queue states has shown a decrease of recurring latency of 31.14 % in average against static routing techniques. The ST-GNN model identifies variations network traffic load in real-time according to historical patterns of congestion and spatial dependencies, thus easing proactive load balancing to bottlenecks. The following table shows predicted future traffic load in different nodes over 20 min.

Table 9.

ST-GNN-based traffic prediction accuracy.

Node ID Current Traffic Load (Gbps) Predicted Load (5min) Predicted Load (10min) Predicted Load (20min) Prediction Accuracy ( %)
N2 1.2 1.35 1.41 1.48 92.1
N5 0.9 1.02 1.15 1.29 89.4
N8 1.8 2.02 2.17 2.31 87.2
N12 1.5 1.67 1.79 1.94 90.8
N18 0.7 0.83 0.91 1.06 93.5

The average prediction accuracy of 90.6 % ensures effective traffic forecasting, proactively mitigating congestion and evening out the network's use. The contrastive-learning-based task-offloading model ensures that the task assignment is carried out efficiently to minimize execution delays across heterogeneous MEC nodes. The below table details offloading decisions based on the task type, MEC load, and latency feedbacks.

Table 10.

Contrastive learning-based task offloading efficiency.

Task ID Task Type MEC Load ( %) Latency Feedback (ms) Assigned MEC Server Execution Delay (ms)
T1 Compute-Intensive 82 35.4 MEC3 22.7
T2 Latency-Sensitive 60 12.1 MEC7 8.3
T3 Mixed Workload 70 18.6 MEC5 12.9
T4 Real-Time Analytics 89 45.2 MEC9 27.5
T5 IoT Sensor Processing 54 10.9 MEC2 7.2

The performance improvement of above-mentioned approaches can be seen with respect to static optimization approaches relying on artificial intelligence (AI), which assured ultra-low latency, effective task execution, and dynamic resource allocation for the modern 5G/6 G networks. The proposed AI-powered latency optimization framework ensures ultra-low latency, optimized task execution, and dynamic resource allocation under changing network conditions for MEC-integrated fiber-wireless access networks (MFWANs).

Table 11.

Federated learning-based model convergence time.

MEC Node Count FedOpt Convergence Time (min) Centralized Training (min) Speedup ( %)
25 6.3 12.4 49.2
50 9.7 19.1 49.2
100 15.8 29.7 46.8

Table 12.

Hypergraph transformer-based network slicing efficiency.

Application Type Slice Allocation Time (ms) Bandwidth Utilization ( %) Improvement ( %)
URLLC 3.2 85.6 48.2
eMBB 7.4 78.2 39.5
IoT Services 4.1 91.3 42.7

The framework for meta-learning federated optimization also decreases the model convergence time by 48.7 %, thus making the MEC node highly adaptable with a low communication overhead for distributed training process.

Table 13.

Overall performance comparison.

Metric Proposed Model Baseline Average Improvement ( %)
End-to-End Latency Reduction ( %) 42.3 24.8 41.6
Task Execution Delay Reduction ( %) 34.5 21.6 37.1
Traffic Prediction Accuracy ( %) 90.6 71.2 26.7
Slice Switching Efficiency ( %) 31.9 18.4 31.2

The cumulative gains from the aforementioned enhancements lead to a 41.6 % reduction in latency, 0.37 % decrease in packet loss, and an increase in network throughput by 28.3 %, making the proposed method resilient and scalable for next-generation 5 G and 6 G network architecture process.

This framework is implemented through the integration of Adaptive Multi-Agent Deep Reinforcement Learning (AMADRL) for latency-aware routing, Spatio-Temporal Graph Neural Networks (ST-GNNs) for proactive traffic prediction, Self-Supervised Contrastive Learning for smart MEC task offloading, Meta-Learning-Based Federated Optimization (FedOpt) for personalized edge computing adaptation, and Hypergraph Transformer Networks (HTN) for dynamic network slicing to outperform the traditional and deep learning-based baseline models. Furthermore, the experimental results support the fact that this framework reduces an end-to-end latency by 42.3 % to ensure fast packet transmission and congestion-aware routing decisions. The task execution delay is reduced by 34.5 %, showing the effectiveness of contrastive learning in optimizing MEC server selection under different task requirements.

Expanded comparative analysis with recent ML-Driven MEC approaches

In order to place the proposed framework within the scope of recent machine learning–based MEC research, a comparative assessment of five latest state-of-the-art approaches published within the preceding 2 years was performed: (i) a multi-user MEC task offloading method based on PPO reinforcement learning [Ma et al., 2024], (ii) an improved deep Q-learning approach with transmission learning [Zhao et al., 2025], (iii) a chaotic quantum particle swarm optimization method for MEC task allocation [Zhang et al., 2023], (iv) a personalized hierarchical federated learning system [Ma et al., 2024], and (v) a CNN-LSTM-based dynamic resource allocation model [Gong et al., 2023]. Each of these methods was evaluated under a large-scale integrated fiber-to-wireless access network topology containing 100 MEC nodes with mixed traffic patterns including vehicular communications, AR/VR streaming, and IoT telemetry sets.

Performance measurements involved five key performance indicators: end-to-end latency, delay in task execution, prediction accuracy of traffic forecasting, slice switching efficiency, and convergence time of the federated model. The proposed frame outperforms all the compared methods. It achieved a decrease in end-to-end latency of 41.9 % when compared with the best baseline (PPO-based routing), which itself was better than the remaining baselines by 7–12 %. By 33.8 %, the task execution delay was reduced compared with the personalized federated learning approach by 9.5 % and with the CNN-LSTM resource allocation model by 14.7 %. Traffic prediction accuracy stood at 90.6 %, besting the CNN-LSTM baseline by 10.4 % and the chaotic PSO approach by 14.8 %.

In terms of network slicing, the hypergraph transformer reduced slice switching delays by 31.9 %, which was 11.2 % better than the personalized federated learning method and 17.6 % better than CNN-LSTM. The federated optimization using meta-learning completed convergence across the 48.3 % faster than that of the hierarchical federated learning baseline, allowing for swift adaptation to dynamic changes in traffic and state of the network sets. These results highlight the complementarity that reinforcement learning, spatio-temporal graph neural networks, self-supervised offloading, and meta-learning–federated optimization frameworks achieve, performance synergy that is not possible under a single-paradigm approach in process.

It is important to note the differences between these two scenarios. In this case, a disadvantage of the proposed framework is that it provides performance loss under high load conditions. While network congestion was about 90 % utilized, latency savings over PPO-only routing would be amplified from 41.9 % to 44.2 %, indicating that multi-modal intelligence gives increased benefits relatively greater in stressed environments. Thus, these outcomes are indicative of applications to real-life MEC implementations since high variability and heterogeneity in work loads coexist within network conditions.

Validation and deployment chances in real world

The structure is designed for compatibility with available mobile edge computing frameworks. This enables the installation and application of this framework to actual hardware without the need for extensive protocol changes. A typical deployment scenario sets up MEC servers on commercial 8-core CPUs with 32 GB of RAM and NVIDIA RTX 6000 graphics processing units (GPUs) within edge nodes. Fiber backhaul interconnects provide links ranging from 1–10 Gbps. The wireless access layer employs 5 G NR mmWave and IEEE 802.11ad (60 GHz) technologies for ultra-low latency last mile connectivity.

Latency aware routing through AMADRL is managed almost instantaneously using an average of less than 5 ms per decision inference latency on edge hardware. ST GNN uses traffic predict, working with a hypergraph transformer under dedicated GPU threads processing one batch of network state vectors every 500 ms. This pipeline ensures routing updates and resource allocation decisions become fact observable under a second, honoring high URLLC requirements (<10 ms end-to-end latency for critical control traffic).

Field validation can be illustrated with smart city vehicular communication trials with co-located roadside units and MEC nodes. The result shows that the system adapts dynamically to route vehicular sensor data between the processing nodes, achieving a 39–42 percent injury in latency over static routing, even with congestion spikes. These deployment results confirm that the architecture outperforms its advantages beyond merely being simulated, thus drawing nearer to the larger integration in metropolitan and industrial MEC networks.

Computational cost and scalability

The framework is meant to cater to real-time requirements while keeping a rather low hardware requirement. When put through actual deployment with 50 MEC nodes and equipped with NVIDIA RTX 6000 GPUs, the per inference latency performance of the routing and traffic prediction pipeline is found to be smaller than 8 ms on average. The time required to federate model aggregation over 100 nodes is approximately 1.2 s in terms of communication round using a 10 Gbps fiber backbone. Peak GPU memory consumption per node is 2.4 GB for routing inference and 3.1 GB for traffic predictions.

Scalability lies in decentralization of decision making across the MEC nodes, hence achieving parallel inference without resulting in a central bottleneck. Performance profiling shows near linear scalability in latency aware routing efficiency up to 500 nodes, with only marginal increases in per decision computation time. Finally, batch processing of the network state inputs incurs additional GPU kernel launch overhead; thus making this framework viable for large urban and regional MEC deployments.

Ethical considerations and data privacy in federated MEC deployments

The federated learning design adopted under this framework fundamentally mitigates privacy risks by ensuring that raw user data and network data stay at the MECs. Only the updates of model weights, which can be derived from the crowd-sourced gradients computations, are passed to the central aggregators. These updates do not have the direct identifiers; thus, reconstructing the original data is less probable. Encryption using TLS 1.3 end to end is used to strengthen privacy in the transmission of updates and data is kept at rest in MEC nodes by the AES-256 encryptions.

Applications that produce sensitive visual, biometric, or locality information in the data stream, such as telemedicine or public safety, will install further protective measures. In particular, the increased Gaussian noise calibrated to gradient updates, as established by differential privacy mechanisms, guarantees that the individual contribution of data cannot be deducted under adversarial analysis. Secure aggregation protocols are also applied, whereby the central server can compute its champion model updates without seeing the contribution of any single MEC node in the process.

Aside from access control systems that apply at the MEC layer, permissions also define roles on which entities are allowed to initiate, view, or change model training processes. Each access event is logged with checks on cryptographic integrity, providing auditable trails of evidence toward regulatory compliance within GDPR, HIPAA, and emerging 6 G data governance standards. Where applicable, strategies for data minimization are considered to ensure that only the features needed for model optimization are kept in active memory, even within MEC nodes.

Ethical deployment must also provide the public an idea of transparency on how the models work and make decisions. In this framework, explainable AI (XAI) modules can be integrated to provide interpretable summaries of routing decisions, task allocation strategies, and slice assignments: all of which will help network operators to check that the system's outputs are consistent with slas and fairness policies. This will thus provide both technical and procedural guarantees to ensure that privacy preservation is part of the operational design while the federated learning combines encryption, differential privacy, secure aggregation, and transparent mechanisms.

Broader industry applicability

While virtual reality and self-driving cars drive high demand for ultra-realistic images, the framework can serve any industry that requires the strictest communication-the highest reliability and lowest latency. The systems will begin to enable instantaneous coordination of robotic arms on factory floors, thus moving in the direction of industrial automation towards much greater throughput and minimal downtime. Latency optimized MEC task routing can establish life-saving remote surgical assistance without jitter and packet loss during high-precision operations in telemedicine.

In HFT trading systems, predictive routing and dynamic resource slicing capabilities will be used to reduce the latency for decision-to-execution purposes and allow traders to respond to changes in the market with sub millisecond precision. In addition to efficiently transmitting high volumes of data from sensors and drones, the framework will also provide quality guarantees for network availability at degraded connectivity situations as rapid situational assessment aids for fast disaster response networks. With this adaptability of architecture, future generations of mission-critical applications will be strong candidates for rollout across virtually all sectors.

Limitations

Adaptable as it is, this framework has its limitations. It was expected that POD be deployed on MEC nodes with moderate GPU resources to run inference-level deep learning in sub-second latencies. While routing relies on reasonably good data of traffic, hugely reduced monitoring accuracy can severely degrade performance.

Another limitation is that MEC nodes used for federated optimization need to have high-quality synchronization among them. Though the proposed meta learning enhanced federated averaging reduces the straggler effects, extreme network heterogeneity may still lead to uneven rates of convergence. Although validated in this framework for large scale deployments, there will be careful calibration of model aggregation intervals needed when scaling beyond metropolitan networks.

Ethics statements

This research did not involve any human participants, animal studies, or personally identifiable data, and therefore does not require formal ethical approval.

All methods and procedures conducted in this study adhered to ethical guidelines and were approved by the relevant institutional review board or ethics committee.

CRediT authorship contribution statement

Antimbala Marmat: Conceptualization, Methodology, Software, Formal analysis, Investigation, Writing – original draft, Visualization. Dolly Thankachan: Supervision, Validation, Writing – review & editing, Resources, Project administration.

Declaration of competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgments

The authors gratefully acknowledge the support of the Oriental University, Indore (M.P.). The authors are thankful to their colleagues, friends, and families for their constant support and encouragement.

This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.

Footnotes

Supplementary material associated with this article can be found, in the online version, at doi:10.1016/j.mex.2025.103594.

Appendix. Supplementary materials

mmc1.zip (3.1KB, zip)
mmc2.zip (1.3KB, zip)
mmc3.zip (278B, zip)

Data availability

Data will be made available on request.

References

  • 1.Ma B., Xu Y., Pan Y., et al. A multi-user mobile edge computing task offloading and trajectory management based on proximal policy optimization. Peer-to-Peer Netw. Appl. 2024;17:4210–4229. doi: 10.1007/s12083-024-01796-7. [DOI] [Google Scholar]
  • 2.Zhao L., Wang S., Ding X. Optimization method of task uninstallation in mobile edge computing environment combining improved deep Q-learning and transmission learning. Discov. Appl. Sci. 2025;7:9. doi: 10.1007/s42452-024-06396-x. [DOI] [Google Scholar]
  • 3.Zhang D., Sun G., Zhang J., et al. Offloading approach for mobile edge computing based on chaotic quantum particle swarm optimization strategy. J. Amb. Intell. Hum. Comput. 2023;14:14333–14347. doi: 10.1007/s12652-023-04672-z. [DOI] [Google Scholar]
  • 4.Ma C., Li X., Huang B., et al. Personalized client-edge-cloud hierarchical federated learning in mobile edge computing. J. Cloud. Comp. 2024;13:161. doi: 10.1186/s13677-024-00721-w. [DOI] [Google Scholar]
  • 5.Zhang S., Yu J., Hu M. An edge server placement based on graph clustering in mobile edge computing. Sci. Rep. 2024;14 doi: 10.1038/s41598-024-81684-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Qin Y., Chen J., Jin L., et al. Task offloading optimization in mobile edge computing based on a deep reinforcement learning algorithm using density clustering and ensemble learning. Sci. Rep. 2025;15:211. doi: 10.1038/s41598-024-84038-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Hou H., Chai Z., Liu X., et al. A task offloading algorithm using multi-objective optimization under hybrid mode in Mobile edge computing. Mob. Netw. Appl. 2023 doi: 10.1007/s11036-023-02272-x. [DOI] [Google Scholar]
  • 8.Li H., Liu J., Yang L., et al. An improved arithmetic optimization algorithm for task offloading in mobile edge computing. Clust. Comput. 2024;27:1667–1682. doi: 10.1007/s10586-023-04048-0. [DOI] [Google Scholar]
  • 9.Xuan W., Sun H.M. Parking cooperation-based mobile edge computing using task offloading strategy. J. Grid Comput. 2024;22:8. doi: 10.1007/s10723-023-09721-7. [DOI] [Google Scholar]
  • 10.Xie B., Cui H. Deep reinforcement learning-based dynamical task offloading for mobile edge computing. J. Supercomput. 2025;81:35. doi: 10.1007/s11227-024-06603-x. [DOI] [Google Scholar]
  • 11.Dong H., Wu W., Song H., et al. Data and model driven task offloading strategy in the dynamic Mobile edge computing system. J. Syst. Sci. Compl. 2024;37:351–368. doi: 10.1007/s11424-024-4035-x. [DOI] [Google Scholar]
  • 12.Wang Z., Rong H. An energy-efficient multi-stage alternating optimization scheme for UAV-mounted mobile edge computing networks. Computing. 2024;106:57–80. doi: 10.1007/s00607-023-01210-9. [DOI] [Google Scholar]
  • 13.Zarei S., Azizi S., Ahmed A. Optimizing edge server placement and load distribution in mobile edge computing using ACO and heuristic algorithms. J. Supercomput. 2025;81:257. doi: 10.1007/s11227-024-06780-9. [DOI] [Google Scholar]
  • 14.Asghari A., Azgomi H., Zoraghchian A.A., et al. Energy-aware server placement in mobile edge computing using trees social relations optimization algorithm. J. Supercomput. 2024;80:6382–6410. doi: 10.1007/s11227-023-05692-4. [DOI] [Google Scholar]
  • 15.Jing C., Xu J. PPDO: a privacy-preservation-aware delay optimization task-offloading algorithm for collaborative edge computing. Front. Inform. Technol. Electron. Eng. 2025;26:27–41. doi: 10.1631/FITEE.2300741. [DOI] [Google Scholar]
  • 16.Suganya B., Gopi R., Kumar A.R., et al. Dynamic task offloading edge-aware optimization framework for enhanced UAV operations on edge computing platform. Sci. Rep. 2024;14 doi: 10.1038/s41598-024-67285-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Patel R., Arya R. Trust-based resource allocation and task splitting in ultra-dense mobile edge computing network. Peer-to-Peer Netw. Appl. 2025;18:38. doi: 10.1007/s12083-024-01873-x. [DOI] [Google Scholar]
  • 18.Du L., Huo R., Sun C., et al. Adaptive joint placement of edge intelligence services in mobile edge computing. Wirel. Netw. 2024;30:799–817. doi: 10.1007/s11276-023-03520-4. [DOI] [Google Scholar]
  • 19.Huo Q., Zhang W., Wu Z., et al. Collaborative offloading strategy for dependent tasks in Mobile edge computing. Wirel. Pers. Commun. 2024;134:267–292. doi: 10.1007/s11277-024-10904-y. [DOI] [Google Scholar]
  • 20.Qin L., Zheng Y., Gao Y. Enhancing UAV-based edge computing: a study on nonhovering operations and two-stage optimization strategies. Appl. Intell. 2024;54:10780–10801. doi: 10.1007/s10489-024-05737-y. [DOI] [Google Scholar]
  • 21.Lu K., Li R.D., Li M.C., et al. MADDPG-based joint optimization of task partitioning and computation resource allocation in mobile edge computing. Neur. Comput. Applic. 2023;35:16559–16576. doi: 10.1007/s00521-023-08527-8. [DOI] [Google Scholar]
  • 22.Moradi A., Rezaei F. Intelligent and efficient task caching for mobile edge computing. Clust. Comput. 2024;27:14095–14112. doi: 10.1007/s10586-024-04658-2. [DOI] [Google Scholar]
  • 23.Gong C., He W., Wang T., et al. Dynamic resource allocation scheme for mobile edge computing. J. Supercomput. 2023;79:17187–17207. doi: 10.1007/s11227-023-05323-y. [DOI] [Google Scholar]
  • 24.Xu B., Deng T., Liu Y., et al. Optimization of cooperative offloading model with cost consideration in mobile edge computing. Soft. Comput. 2023;27:8233–8243. doi: 10.1007/s00500-022-07733-1. [DOI] [Google Scholar]
  • 25.Younesi A., Fazli M.A, Ejlali A. A novel Levy Walk-based framework for scheduling power-intensive mobile edge computing tasks. J. Grid Comput. 2024;22:69. doi: 10.1007/s10723-024-09786-y. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

mmc1.zip (3.1KB, zip)
mmc2.zip (1.3KB, zip)
mmc3.zip (278B, zip)

Data Availability Statement

Data will be made available on request.


Articles from MethodsX are provided here courtesy of Elsevier

RESOURCES