Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2025 Dec 1;15:42875. doi: 10.1038/s41598-025-27120-8

Recognition method for microgrid switching misoperations based on multi-source heterogeneous data fusion

Hongzhao Yang 1,3,4,, Zhan Zhang 1,4, Shijie Zhang 2,3, Rui Liang 2,3
PMCID: PMC12669576  PMID: 41326441

Abstract

To enhance the safety of microgrid switching and the identification of misoperations, we propose Time-Synchronized Misoperation Recognition (TS-MR), a method tailored to switching operations. The approach performs rule-based pre-screening grounded in operating procedures and anti-misoperation interlocking, achieves millisecond-level time synchronization of multi-source heterogeneous data via a two-stage scheme that combines variational Bayesian inference with a UKF, and employs a fusion of a Transformer, a TCN, and a GNN for cross-modal representation with interpretable discrimination. Laboratory records constitute the training, validation, and test sets; HIL data are used solely for independent cross-validation; and public datasets are used only for cross-domain robustness calibration, and none contributes to training, validation, or threshold tuning. Under a unified evaluation protocol, TS-MR attains 94.69% accuracy and an AUC of 0.977 in typical switching scenarios; end-to-end latency is about 80 ms; the core forward-pass latency is about 42 ms; the Inline graphic per-inference latency is 55 ms; and computational complexity is about 3.4 GFLOPs. Compared with CNN-BiLSTM, ConvLSTM, and GAT under identical preprocessing, time synchronization, and fixed random seeds, TS-MR improves accuracy by 0.9 to 3.7% points and AUC by 0.024 to 0.057. These results indicate that TS-MR provides high-confidence misoperation recognition and interpretable assessment for microgrid switching while satisfying engineering-grade real-time constraints.

Supplementary Information

The online version contains supplementary material available at 10.1038/s41598-025-27120-8.

Keywords: Microgrid switching, Misoperation recognition, Time synchronization, Multi-source heterogeneous data fusion, Graph neural network, Anti-misoperation interlocking

Subject terms: Energy science and technology, Engineering, Mathematics and computing

Introduction

With the large-scale integration of renewable energy, microgrids exhibit stronger state coupling and higher uncertainty during critical operations such as grid-connected and islanded switching, topology reconfiguration, energy-storage coordination, and load transfer1. Deviations in the operating sequence, interlocking logic, or state recognition can trigger misoperations, leading to voltage excursions, frequency oscillations, and relay malfunctions; in severe cases, cascading outages may ensue2,3. In practice, the combined effects of communication jitter, network latency, meteorological disturbances, and equipment failures make cross-source time mismatch a major source of false judgments. Experimental and replay studies indicate that when multi-source time-synchronization errors reach the millisecond scale, event-level recognition performance degrades sharply4,5. Consequently, traditional mechanical interlocks, electrical blocking, and manual procedural verification are insufficient for data-intensive, disturbance-prone microgrids that are increasingly instrumented with edge intelligence6,7.

Existing research has advanced along two main directions. (i) Rule–data collaboration. Operating procedures and interlocking logic are formalized as non-negotiable safety boundaries to rapidly remove explicit violation sequences. In parallel, probabilistic risk priors are introduced to characterize environmental uncertainty and operational risk intensity; their role is confined to weighting during training and threshold calibration and does not replace rule judgments or interlocking logic, preserving both auditability and adaptability8,9. (ii) Two-stage time synchronization. Engineering practice commonly employs IEEE 1588 Precision Time Protocol (PTP) and timestamping based on phasor measurement units (PMUs); however, when jitter, packet loss, and sampling-rate mismatch coexist, single-stage synchronization leaves residual bias10,11. A two-stage strategy is therefore adopted: fixed or slowly varying offsets are first removed, and then time-varying residuals are compensated using variational Bayesian inference and an Unscented Kalman Filter (UKF). Synchronization accuracy and robustness are summarized with quantile statistics to capture central tendency and tail risk, alongside magnitude comparisons to standard synchronization references for engineering interpretability12,13.

On representation and discrimination, early single-modality time-series models (for example, hybrids of CNN and BiLSTM, or a convolutional network coupled with an LSTM) showed some robustness for power-event detection but remained limited in capturing semantics across modalities and long-range dependencies14. Attention-centric multimodal frameworks have since emerged: the Transformer aligns semantics across modalities and weights features; the Temporal Convolutional Network (TCN) efficiently models sequential dynamics; and the Graph Neural Network (GNN) explicitly encodes equipment and circuit topology, strengthening causal association and localization15,16. Recent work tends to fuse these approaches to form a unified representation of semantics, temporal patterns, and topology, together with interpretability tools such as attention heatmaps, permutation importance, and checks for evidence consistency, thereby improving auditability and traceability for operations and maintenance17,18.

Regarding data and evaluation transparency, best practices favor real and laboratory measurements as the primary data, supplemented by hardware-in-the-loop (HIL) and simulated data to cover boundary conditions, with open datasets for cross-domain robustness calibration. It is advisable to disclose, by partition, each subset’s source, proportion, time window, sampling rate, and labeling protocol; to define training, validation, and test splits that preclude leakage; and to report metrics separately for real versus augmented or public subsets to ensure extrapolatability and reproducibility1921. At the engineering-deployment level, collaboration between the cloud and the edge is increasingly prevalent: the edge executes low-latency, real-time inference, while the cloud handles retraining, evaluation, and model distribution. Evaluation typically reports computational complexity (GFLOPs), end-to-end latency, and quantile-based stability, guiding model compression, distillation, and operator-level optimization2224. In parallel, data-driven cyber-protection has progressed (for example, false-data-injection detection and localization in DC microgrids using subspace identification and residual generators), but its scope—cyberattacks and data tampering—differs from the switching-misoperation task addressed here, which hinges on multi-source time synchronization and procedural-logic embedding25.

To address these challenges, we propose Time-Synchronized Misoperation Recognition (TS-MR), following the principle of rule-driven filtering, two-stage time synchronization, and multimodal cooperative discrimination. Procedural and interlocking logic provide deterministic screening; time synchronization uses a dual-stage scheme that combines variational Bayesian inference with a UKF to suppress cross-domain timestamp bias, with robustness summarized by quantile statistics and compared in magnitude with industrial synchronization references. The discrimination layer integrates the complementary strengths of the Transformer, TCN, and GNN to jointly model contextual semantics and equipment topology. To avoid replacing deterministic safety rules with statistical risk, a risk module based on Monte Carlo simulation is used only as a training-phase prior for sample weighting and threshold scheduling and is excluded from rule or interlock compliance decisions.

Contributions

  1. Decoupled rule filtering and probabilistic priors. We build a rule-driven anti-misoperation and interlocking module and restrict Monte Carlo usage to training-phase risk priors for sample weighting and threshold scheduling, ensuring independence from deterministic rule judgments.

  2. Two-stage time synchronization with engineering comparators. A variational Bayesian and UKF scheme achieves millisecond-level multi-source synchronization; we report alignment-error quantiles (central tendency and tail risk) and provide magnitude comparisons against IEEE 1588 (PTP) and PMU-based references.

  3. Multimodal fusion with interpretability. A collaborative framework that integrates the Transformer, TCN, and GNN unifies semantic information and topology; attention heatmaps, permutation importance, and node-contribution scores support operator-level interpretation and traceability.

  4. Partitioned data and a unified evaluation protocol. We curate a dataset with real and laboratory measurements as the main body and public data for augmentation; we report results separately for real versus augmented subsets and adopt a reproducible protocol that discloses dataset versions, time windows, preprocessing and synchronization workflows, and random seeds, while reporting efficiency indicators, including GFLOPs and end-to-end latency, for comparability.

Mathematical modelling

TS-MR comprises four stages: (i) rule-driven pre-screening, (ii) two-stage time synchronization, (iii) multimodal representation and fusion, and (iv) decision and constrained optimization. The overall workflow is shown in Fig. 1; the solution workflow is illustrated in Fig. 2.

  1. Rule-driven pre-screening: Using interlocking logic, authorization consistency, temporal safety margins, and load and communication boundaries, this stage establishes verifiable operational constraints and risk descriptors. It directly outputs three labels—Pass, Violation, Review—and preserves traceable evidence. Hydrogen-related anomalies are incorporated into the risk channel as indicators with weights to quantify their impact on equipment stability and multi-energy coupling.

  2. Two-stage time synchronization: A global mean drift is first removed, followed by dynamic residual compensation to suppress desynchronization caused by switching transients; a synchronization confidence measure is also provided. Magnitude comparisons with IEEE 1588 and GPS/UTC-disciplined PMU synchronization are reported in the Methods and Results as engineering references rather than replacements for industrial time synchronization.

  3. Multimodal representation and fusion: Cross-modal attention with a Transformer aligns and weights information; TCNs extract the dynamics of operational sequences; GNNs encode equipment topology and interlocking constraints. The fused representation supports discrimination and yields interpretable Top-K evidence (for example, attention heatmaps and node-contribution scores) to facilitate operations and maintenance verification and traceability.

  4. Decision and constrained optimization: Within the feasible domain defined by rule-consistency and synchronization-consistency constraints, the method enforces budgets on end-to-end latency and computational complexity (GFLOPs) and outputs the misoperation probability and risk level, balancing accuracy and robustness. The Monte Carlo risk module is used only as a training-phase prior for risk weighting and sample reweighting; it does not participate in rule screening or real-time compliance decisions.

Fig. 1.

Fig. 1

TS-MR architecture. Stage (1) rule-driven pre-screening; Stage (2) two-stage time synchronization (mean-bias calibration and dynamic residual compensation); Stage (3) multimodal representation and fusion (Transformer, TCN, GNN); Stage (4) decision and optimization under feasibility, rule-consistency, and synchronization-consistency constraints with latency and GFLOPs budgets. Monte Carlo-based risk assessment is used only as a training prior.

Fig. 2.

Fig. 2

TS-MR solution workflow. The left column shows constraint–feature modeling (multi-source); the right column shows the PGD-based optimization. The outputs are the misoperation probability Inline graphic and the corresponding risk level Inline graphic. Model coefficients are calibrated offline using a hold-out validation and k-fold cross-validation protocol; see Supplementary Table S2.

Rule-driven anti-misoperation model

To quantify the rule-compliance risk of switching commands, five components are defined: equipment state, operational sequence, voltage–communication coordination, assembly and disassembly compliance, and multi-energy coupling. The overall objective function is

graphic file with name d33e380.gif 1

where Inline graphic denotes the overall rule-compliance risk; Inline graphic the Inline graphic-th risk component; and Inline graphic a non-negative weight calibrated from historical statistics and operation-maintenance experience.

  1. Equipment-state risk
    graphic file with name d33e409.gif 2
    where Inline graphic is the equipment-state risk; Inline graphic is the Inline graphic-th sub-risk item and Inline graphic its weight; Inline graphic is the expected sampling or sequencing jitter; Inline graphic is the weight of the cloud–edge interaction cost Inline graphic; Inline graphic is the hydrogen-energy anomaly-intensity index; and Inline graphic is its influence weight within the equipment-state channel.
  2. Operational-sequence risk
    graphic file with name d33e456.gif 3
    where Inline graphic is the operation-sequencing risk; Inline graphic is the weight of sub-component Inline graphic; Inline graphic is the Inline graphic-th operation-sequencing sub-risk, reflecting the deviation from the prescribed operation order; Inline graphic is the disturbance term of the control chain; and Inline graphic is the influence weight of hydrogen anomalies in the sequence-risk component.
  3. Voltage–communication coordination risk
    graphic file with name d33e495.gif 4
    where Inline graphic is the integrated risk of voltage–communication coordination; Inline graphic denotes the risk of voltage or frequency limit violation; Inline graphic the risk induced by delay and jitter in the communication chain; and Inline graphic the transformer–breaker coordination risk. The coefficients Inline graphic,Inline graphic,Inline graphic weight the respective sub-risks.
  4. Assembly and disassembly procedure–compliance risk
    graphic file with name d33e534.gif 5
    where Inline graphic is the compliance risk of assembly and disassembly procedures; Inline graphic is the Inline graphic-th checklist sub-risk; Inline graphic is the corresponding item weight; Inline graphic denotes the compliance-related disturbance; Inline graphic is the influence weight of the hydrogen anomaly within the compliance risk; and Inline graphic denotes a generic disturbance term.
  5. Multi-energy coupling risk
    graphic file with name d33e573.gif 6
    where Inline graphic is the multi-energy coupling risk; Inline graphic is the set of energy types; Inline graphic is the availability probability of energy type Inline graphic at node Inline graphic; Inline graphic is the node weight; and Inline graphic represents the uncertainty induced by power surges or output fluctuations.
  6. Hydrogen-anomaly propagation mechanism
    graphic file with name d33e612.gif 7
    where Inline graphic is the Sigmoid function; Inline graphic and Inline graphic are hyperparameters fitted on the validation set to ensure that Inline graphic decreases monotonically as Inline graphic increases. Here Inline graphic is the hydrogen anomaly intensity index, obtained by normalizing and weighting multi-signal evidence after time synchronization. Relevant evidence includes leakage or concentration alarms, abrupt drops in hydrogen-storage pressure, reduced or abnormally falling PEM fuel-cell stack voltage, electrolyzer trips, and large power fluctuations. As Inline graphic rises, Inline graphic decreases, thereby reducing Inline graphic in (6) and increasing Inline graphic, reflecting the adverse impact of hydrogen anomalies on system stability.
  7. Rule and engineering constraints

    1. Temporal constraint
      graphic file with name d33e671.gif 8
      where Inline graphic and Inline graphic are the off and on switching instants of device Inline graphic at operation step Inline graphic; Inline graphic is the minimum safe interval between steps Inline graphic and Inline graphic; Inline graphic denotes the operation uncertainty of device Inline graphic; Inline graphic is the set of all devices; and Inline graphic is the set of operation steps.
  8. Interlocking constraint
    graphic file with name d33e727.gif 9
    where Inline graphic is the binary state of circuit breaker Inline graphic (0 for open, 1 for closed); Inline graphic is the interlocking constraint associated with step Inline graphic; Inline graphic is the interlock-risk weight; Inline graphic is the interlock threshold bias; Inline graphic is the hydrogen-anomaly weight within the interlocking constraint; and Inline graphic is the permissible interlock tolerance.
  9. Load constraint
    graphic file with name d33e770.gif 10
    where Inline graphic is the real-time load of device Inline graphic; Inline graphic is the rated load limit of device Inline graphic; Inline graphic is the load-disturbance term; and Inline graphic is the hydrogen-anomaly weight for voltage–load coordination scenarios.
  10. Time-synchronization deviation constraint
    graphic file with name d33e805.gif 11
    where Inline graphic is the synchronization deviation at time Inline graphic; Inline graphic is the allowable upper bound at time Inline graphic; Inline graphic is the error-compensation term at time Inline graphic; and Inline graphic is the hydrogen-related time-synchronization weight.
  11. Feature-consistency constraint
    graphic file with name d33e844.gif 12
    where Inline graphic is the feature-space distance between Inline graphic and Inline graphic; Inline graphic is the difference in the Inline graphic-th feature dimension; Inline graphic is the feature-consistency threshold; Inline graphic denotes the threshold disturbance term, reflecting the uncertainty in threshold setting. Inline graphic is the expansion disturbance for feature-consistency deviation; and Inline graphic is the hydrogen-anomaly weight within the feature-consistency constraint.
  12. Historical-consistency constraint
    graphic file with name d33e892.gif 13
    where Inline graphic is the execution time of the current operation; Inline graphic and Inline graphic are the historical mean and standard deviation; Inline graphic is the tolerance coefficient for historical deviation; Inline graphic is the disturbance term associated with Inline graphic; and Inline graphic is the hydrogen-related weight under the historical-consistency scenario.
  13. Assembly and disassembly authorization constraint

    graphic file with name d33e931.gif 14
    where Inline graphic is the authorization requirement for the Inline graphic-th operation; Inline graphic is the authorization record for the Inline graphic-th operation; Inline graphic is the global safety-margin disturbance; Inline graphic is the hydrogen-energy weight under the authorization scenario; and Inline graphic denotes the disturbance due to misjudgment or repeated authorization.
  14. Complexity and real-time constraint
    graphic file with name d33e970.gif 15
    where Inline graphic is the per-inference FLOPs of the model at parameter set Inline graphic; Inline graphic is the upper bound of allowable complexity; Inline graphic is the end-to-end inference latency; and Inline graphic is the latency limit; Inline graphic denotes the model parameter set. These bounds serve as engineering-level constraints to guide deployment–performance matching; they are set with reference to engineering practice in microgrid control and grid interconnection (e.g., IEEE 2030.7 and IEEE 1547) and are used solely for benchmarking, not as compliance declarations.26,27

Two-stage time synchronization

During switching operations, discrepancies in sampling frequency and communication paths across SCADA, PMU, video, voice, and electronic operation tickets introduce temporal deviations of approximately 100–500 ms28. To obtain a unified timestamp across modalities, a two-stage time synchronization mechanism is adopted. In the first stage, mean bias is eliminated via statistical averaging; in the second stage, dynamic residuals are compensated by combining Kalman filtering (UKF where applicable) with variational Bayesian inference. The entire synchronization process follows the precision-time specifications of IEEE 1588, IEC 61850–9-3, and GPS-PMU, meeting engineering-grade consistency requirements2931.

Before executing the two stages, an adaptive gating coefficient is defined to modulate the correction strength and to provide robust regularization information for the second stage:

graphic file with name d33e1024.gif 16

where Inline graphic is the adaptive correction strength at time Inline graphic; Inline graphic is the residual-correction weight; Inline graphic is the instantaneous deviation of the multi-source signals with respect to the PMU at time Inline graphic; Inline graphic is the jitter-error correction term; Inline graphic is the system-state indicator at time Inline graphic; Inline graphic is the Monte Carlo regularization weight; and Inline graphic is the Monte-Carlo-based robust regularizer used to enhance stability rather than for rule screening.

Stage 1: Mean-bias correction

graphic file with name d33e1073.gif 17

where Inline graphic is the average time offset of channel Inline graphic; Inline graphic and Inline graphic are the raw and PMU timestamps of the Inline graphic-th sample, respectively; Inline graphic is the total number of samples; and Inline graphic denotes the unified timestamp (for channel Inline graphic, sample Inline graphic ) after mean-bias correction.

Stage 2: Dynamic residual compensation

graphic file with name d33e1117.gif 18

where Inline graphic is the synchronized timestamp of channel Inline graphic at sample Inline graphic; Inline graphic is the AR(1) residual-model coefficient; Inline graphic is the previous-step residual estimate; Inline graphic is the prior covariance; Inline graphic and Inline graphic are the process-noise and observation-noise covariance matrices; Inline graphic is the current observation; and Inline graphic is the compensation term for time jitter and sampling quantization error.

Multimodal representation and fusion

graphic file with name d33e1166.gif 19

where Inline graphic denotes the fused feature representation at layer Inline graphic; Inline graphic is the cross-node attention weight, measuring the contribution of node Inline graphic to node Inline graphic; Inline graphic and Inline graphic are the feature vectors of nodes Inline graphic and Inline graphic at layer Inline graphic; Inline graphic is the risk-label weight for class Inline graphic; Inline graphic is the indicator matrix for class Inline graphic; Inline graphic is the projection weight matrix of layer Inline graphic; Inline graphic is the bias vector; Inline graphic represents scenario-level uncertainty disturbance in multimodal settings; Inline graphic is the hydrogen-scenario fusion factor; and Inline graphic denotes the multi-scenario robustness expectation of the Monte Carlo term.

Model decision and optimization

This section completes the closed loop from probabilistic decision to constrained optimization. First, the misoperation probability is obtained by Eq. (20) to form a binary decision; then the global objective within the feasible engineering domain (Eqs. (22)–(23)) is solved using Projected Gradient Descent (PGD)32.

  1. Probabilistic decision

    graphic file with name d33e1279.gif 20
    where Inline graphic is the Sigmoid activation; Inline graphic is the attention weight; Inline graphic,Inline graphic are the feature vectors of nodes Inline graphic and Inline graphic ; Inline graphic is the bias; Inline graphic is the time-weighted disturbance; Inline graphic is the misoperation-risk factor under hydrogen scenarios; Inline graphic is the Monte Carlo–based scene-verification term; and Inline graphic is the generalized error term.
    graphic file with name d33e1330.gif 21
    where Inline graphic is the risk level of node Inline graphic; Inline graphic is the grade index; Inline graphic is the total number of grades; and Inline graphic are the probability thresholds.
  2. PGD-based constrained optimization
    graphic file with name d33e1360.gif 22
    where Inline graphic is the total loss; Inline graphic the classification loss; Inline graphic the Frobenius-norm regularization (Inline graphic); Inline graphic,Inline graphic,Inline graphic,Inline graphic are the weights of the corresponding terms; Inline graphic is the compliance vector computed from Eqs. (8)–(14); Inline graphic is the rule-consistency loss; Inline graphic the time-synchronization penalty; Inline graphic the MCRA-prior regularizer; and Inline graphic the cloud-resource penalty.
    graphic file with name d33e1425.gif 23
    where Inline graphic is the feasible engineering domain jointly defined by Eqs. (8)–(15); Inline graphic is the validation accuracy and Inline graphic its lower bound. To avoid interference with relay protection and in-station communication, the upper bound of inference latency is set to the engineering-grade deployment limit Inline graphic (batch = 1). Computational complexity is bounded by Inline graphic. The threshold Inline graphic is determined from the ROC operating point (Inline graphic, Inline graphic, F1-max). Regularization and latency limits reference IEC 61,850–5 and IEEE 1547–2018 time scales.

PGD-based optimization procedure

  1. Initialization: Set Inline graphic, learning rate Inline graphic, and iteration index Inline graphic.

  2. Gradient update
    graphic file with name d33e1498.gif 24
    where Inline graphic is the unprojected update; Inline graphic the parameters at iteration Inline graphic; Inline graphic the learning rate; Inline graphic the gradient; and Inline graphic the update-noise disturbance.
  3. Projection mapping
    graphic file with name d33e1533.gif 25
    where Inline graphic is the projection operator onto Inline graphic; Inline graphic is the projection-uncertainty disturbance.
  4. Learning-rate adjustment
    graphic file with name d33e1555.gif 26
    where Inline graphic is the learning rate at iteration Inline graphic; Inline graphic,Inline graphic are its bounds; Inline graphic is the annealing period; Inline graphic the adjustment-noise term; Inline graphic the hydrogen-scenario learning-rate correction; and Inline graphic the multi-scenario consistency term.
  5. Early stop: If validation accuracy Inline graphic shows no improvement for Inline graphic consecutive windows and Inline graphic, terminate optimization and output the optimal parameter set Inline graphic.

  6. Output. Using Inline graphic for inference, obtain the misoperation probability Inline graphic and the corresponding risk level Inline graphic for node Inline graphic(see Eqs. (20)–(21)). See Table1 for an end-to-end example.

Table 1.

Example of end-to-end calculation for device-state risk.

Step Symbol Description Example value Corresponding Eq
Step-1 Post-synchronization analysis window (the period after unified time alignment by Eqs. (16)–(18)), actual event window 1 s (16)–(18)
Step-2 Inline graphic Sub-risk 1: load deviation (computed by Eq. (10), inserted into Eq. (2)) 0.067 (2)
Step-3 Inline graphic Sub-risk 2: position inconsistency (computed by Eq. (12), inserted into Eq. (2)) 0.020 (2)
Step-4 Inline graphic Sub-risk 3: interlock deviation (computed by Eq. (9), inserted into Eq. (2)) 0.015 (2)
Step-5 Inline graphic Weights of the three sub-risks (normalized so that Inline graphic) Inline graphic (2)
Step-6 Inline graphic Expected value of sampling and timing jitter 0.0015 (2)
Step-7 Inline graphic Cloud–edge interaction term 0.0020 (2)
Step-8 Inline graphic Impact of hydrogen anomalies on device-state risk (modeled and inserted via Eqs. (2) and (7)) 0.0040 (2),(7)
Step-9 Inline graphic Aggregated device-state risk (combined by Eq. (2)) 0.0505 (2)
Step-10 Inline graphic Weight of each risk type; here Inline graphic 0.40 (1)
Step-11 Inline graphic,Inline graphic,Inline graphic,Inline graphic Remaining risk categories (this window not triggered) 0 (3)–(6)
Step-12 Inline graphic Aggregated rule-level risk (combined by Eq. (1)) 0.0202 (1)
Step-13 Engineering-grade thresholds (Pass/Review/Block) Pass < 0.010; Review 0.010–0.030; Block > 0.030

End-to-end worked example

Experimental setup and datasets

Experimental environment

Computational experiments were run on an Intel Xeon Gold 6230 processor (20 cores, 2.1 GHz) with 128 GB RAM and an NVIDIA RTX A6000 GPU (48 GB VRAM). Simulation modeling was performed in MATLAB R2024b, and data processing and statistical analysis were conducted in Python 3.9 with PyTorch 2.0.1. At the software layer, an IEEE 1588 Precision Time Protocol (PTP) time synchronization link was established using ptpd v2.3.1 to evaluate time alignment and quantify synchronization errors; the resulting comparisons are provided as engineering references only and do not constitute a standards-compliance statement33,34. To ensure reproducibility, the random seed was fixed at 12,345.

The physical platform is the distribution-network prototype system of the National Key Laboratory for Power Grid Disaster Prevention and Mitigation at Changsha University of Science and Technology. It adopts a single-bus architecture with approximately 20 grid-connected units (photovoltaics, wind generation, hydrogen storage, electrochemical storage, and typical loads). The system voltage level is 10 kV with a 0.63 MVA transformer; the total grid-connected rated capacity is about 300 kW, including PV 60 kW, WT 30 kW, BESS 150 kW, and H2 60 kW. PMUs and SCADA are deployed at major interconnection points and feeder outlets for phasor, frequency, and status acquisition as well as event alignment. Communication and time-synchronization interfaces include Modbus-TCP, IEC 61,850-GOOSE, and PTP .

Data sources and uses

This study employs three categories of data: laboratory real data for model training, validation, and testing; HIL data for independent cross-validation and robustness enhancement; and public datasets for cross-domain robustness calibration and distribution comparison. The composition and roles of each category are summarized in Table 2.

Table 2.

Data sources and roles.

Mode Records(n) Sampling rate and time coverage Key features and description Role in this study
SCADA and PMU 32,000 SCADA: 1 Hz (± 120 ms jitter); PMU: 60 sps; coverage: 2025–09-03 08:00–16:53 (approximately 8.9 h); both streams resampled to 1 Hz after the two-stage time-synchronization pipeline Voltage and current phasors (magnitude and angle), system frequency, active and reactive power, switching status, and other operational quantities Primary data source for model training, validation, and testing
Operation tickets 3200 Event-level; coverage: 2025–09-03 08:00–2025–09-04 07:59 Operation tasks, step sequence, device labels, scheduled actions, execution time, and authorization levels Operation tasks, step sequence, device labels, scheduled actions, execution time, and authorization levels
Operator command speech 3200 Event-level; original 16 kHz audio; only event timestamps and recognition confidence retained Voice commands and operator intent; trigger words and confidence scores Used for semantic intent recognition and multimodal verification
Video events 3200 Event-level; original 30 fps video; only event timestamps and scene summaries recorded Operator actions and scene context; visual events and contextual summaries Used for visual-assisted verification and scene-consistency analysis
Environment and working conditions 288 5-min sampling; coverage: 2025–09-03 08:00–2025–09-04 07:55 Irradiance, wind speed, temperature, airflow, and load-monitoring variables of the external environment Used for operational-boundary modeling and external-disturbance analysis
Risk labels 3200 Label-level; coverage: 2025–09-03 08:00–2025–09-04 07:59 Misoperation categories and risk levels (high, medium, and low) Used for model supervision and evaluation benchmarking
  1. Laboratory real data

    Data were collected from the laboratory microgrid platform and include SCADA and PMU records together with paired operation tickets, video, and voice events. The core SCADA and PMU window comprises 32 000 samples at 1 Hz with approximately ± 120 ms jitter, covering 2025–09-03 08:00 to 16:53; other modalities extend to 2025–09-04 07:59. All data were aligned under unified PTP or GPS/UTC time synchronization using a two-stage alignment mechanism. The original PMU signals (60 fps) were aligned to 1 Hz, and the SCADA 1 Hz sequence served as the sampling index. Video and voice streams retain only event timestamps and confidence scores. This subset constitutes the primary source for model training, validation, and testing.

  2. Hardware-in-the-loop data

    Generated on the StarSim-HIL platform, this subset provides approximately 9 100 multi-modal records (about 20%) and is used solely for independent cross-validation and robustness enhancement.

  3. Public datasets

    Two open datasets were used: the NREL Hydrogen Microgrid Dataset v1.2 (containing hydrogen operation data) and “Power-Data-from-Microgrid” (Kaggle, v2024, including power, voltage, frequency, temperature, and other microgrid/IoT sensor data), totaling about 4 600 records (approximately 10%)35,36. These data are used only for cross-domain robustness calibration and distribution comparison and are excluded from training, validation, and testing. No SMOTE is applied to the public datasets.

Statistical characteristics and distribution comparison

After temporal windowing, the main electrical quantities are summarized by mean, standard deviation, minimum, and maximum in Table 3. Distribution differences across laboratory real data, HIL data, and public datasets are compared in Table 4.

Table 3.

Summary statistics of key electrical variables (windowed samples, n = 32 000).

Feature Mean Standard deviation Minimum Maximum
Voltage magnitude Inline graphic(p.u.) 0.9998 0.0198 0.9709 1.0367
Current magnitude Inline graphic(p.u.) 0.499 0.288 0.1534 0.989
Frequency Inline graphic(Hz) 60.029 0.0502 59.875 60.105
Active power Inline graphic(kW) 50.175 28.890 0.002 99.996
Reactive power Inline graphic(kvar)  − 0.071 28.936  − 49.999 50.000

Table 4.

Comparative statistics across data sources.

Data type Mean voltage (p.u.) Voltage standard deviation (p.u.) Frequency range (Hz) Average hydrogen flow rate
Laboratory real data 0.9998 0.32 59.875–60.105 30.07
Hardware-in-the-loop (HIL) data 1.001 0.32 59.88–60.10 28.85
Public dataset 1.005 0.35 59.70–60.30 31.02

Preprocessing workflow and class imbalance handling

  1. Data cleaning. Outliers are removed using the three-sigma criterion together with scenario rules. Windows with a missing-value ratio ≥ 1% are discarded; remaining gaps are repaired by forward filling and linear interpolation.

  2. Time synchronization. PMU timestamps serve as the unified time baseline. A two-stage mechanism (VB followed by UKF) is employed. On the validation set, the mean synchronization error is 1.8 ms, the standard deviation is 0.5 ms, and Inline graphic is 3.0 ms. Comparative evaluation is conducted against PTP and GPS-PMU references.

  3. Feature engineering. Electrical quantities are standardized; video streams provide event counts and confidence; speech yields keywords and ASR confidence; operation tickets are parsed into operation sequences and authorization levels; environmental variables retain their original engineering units.

  4. Windowing and alignment. Event-centered synchronized slicing is applied with a fixed window Inline graphic and step Inline graphic.

  5. Partitioning and balancing. Laboratory real data are split 8:1:1 into training, validation, and test sets. SMOTE is applied only within the training set to balance minority classes. Public/open data are used solely for cross-domain robustness calibration and distribution comparison, and do not participate in main training, validation, or threshold setting.

Time-synchronization parameters and stopping criteria

A two-stage VB + UKF synchronization scheme is adopted with the following settings.

  1. VB stage (mean-offset correction): Convergence threshold Inline graphic; maximum 100 iterations; initial variance Inline graphic. Process and measurement noise variances are selected by grid search: Inline graphic,Inline graphic.

  2. UKF stage (dynamic residual compensation): Unscented-transform parameters: Inline graphic, Inline graphic. Sliding-window length Inline graphic; the residual purification threshold is adaptively tuned.

  3. Global stopping rule: Training stops when the change in the mean residual over five consecutive windows satisfies Inline graphic and the 95th-percentile error Inline graphic shows no further decrease.

HIL connectivity and interface verification

To evaluate the deployability of TS-MR under engineering-analogous conditions, a StarSim-HIL testbed was constructed to establish bidirectional coupling with the compute side, forming a hardware-in-the-loop verification environment (Fig. 3). The experimental platform comprises FPGA-based acquisition, timing-control, and host-controller units. The entire link adopts IEEE 1588 for a unified time base, and interfaces support Modbus-TCP and IEC 61,850-GOOSE.

Fig. 3.

Fig. 3

StarSim-HIL testbed and interfaces.

I/O capabilities: 16 analog channels (1 MS/s, 16-bit, ± 10 V) and 32 digital channels (5 V TTL and ± 25 V). Test scripts cover scenarios such as forced interlock override, grounding-conductor removal, and unauthorized entry into an energized interval. During testing, I/O waveforms, event timestamps, and synchronization step offsets are recorded to verify connectivity and temporal-sequence consistency.

Experimental validation and analysis

Network architecture and training configuration

TS-MR adopts a multi-branch encoding–fusion–classification architecture. The TCN extracts temporal features from SCADA and PMU; the Transformer performs cross-modal semantic fusion; and the GNN models topology-aware interactions among equipment nodes. After feature extraction and concatenation, the fused representation is flattened by an MLP and passed to a softmax layer for class-weighted probability prediction. The baseline models (CNN-BiLSTM, ConvLSTM, and GAT) are trained under a unified search strategy with minor hyperparameter adjustments to ensure fair comparability. Key network and training parameters are summarized in Table 5. The average end-to-end inference latency per batch (batch = 1) is 80 ms (including input normalization and feature concatenation) . The core forward time is 42 ms (Inline graphic is approximately 55 ms), and the computational complexity is 3.4 GFLOPs, consistent with Table 9 and used for engineering-grade benchmarking.

Table 5.

TS-MR network architecture and training configuration summary.

Parameter category Parameter name Value
Network architecture Transformer layers 4
Attention heads 8
TCN convolutional layers 4
GNN layers 3
Hidden dimension 128
Training configuration Learning rate 0.001
Batch size 64
Epochs 50
Optimizer Adam
Dropout ratio 0.2
Efficiency metrics Core forward latency (mean and Inline graphic) 42 ms and 55 ms
End-to-end inference latency (batch = 1) 80 ms
Peak memory usage 2.8 GB
Computational complexity (GFLOPs) 3.4

Table 9.

Comprehensive performance and efficiency comparison under the unified evaluation protocol.

Model Accuracy (%) Precision (%) Recall (%) F1 (%) AUC End-to-end latency (ms) GFLOPs
TS-MR 94.69 92.5 95.1 93.8 0.977 80 3.4
CNN-BiLSTM 93.8 92.4 93.2 92.8 0.953 108 2.1
ConvLSTM 92.3 91.2 91.8 91.5 0.941 102 2.7
GAT 91.0 89.8 90.5 90.1 0.920 92 1.4

Rule consistency and risk suppression

  1. Risk distribution suppression after rule embedding

    According to the risk terms defined in Eqs. (1)–(7), the test-set tail distribution shifts left after applying the rule-consistency mechanism. The 95th-percentile risk score decreases from 0.65 to 0.55, indicating effective suppression of high-tail risks and mitigation of low-probability hazards (Fig. 4).

  2. Classification improvement under a unified threshold.

    Using a fixed threshold derived from the “pre-rule” ROC curve at TPR = 0.90, overall accuracy increases from 87.1% to 92.5% after applying the rule constraints, with both the false positive rate (FPR) and the false negative rate (FNR) reduced. A McNemar test shows the improvement is statistically significant (Inline graphic). Detailed results are provided in Table 6.

    Table 7.

    Per-class metrics for typical misoperation scenarios.
    Operation type Precision (%) Recall (%) F1 (%) AUC
    Unauthorized grounding disconnection 93.2 95.1 94.1 0.985
    Forced interlock override 92.5 94.8 93.6 0.987
    Unauthorized command execution 94.1 93.2 93.6 0.963
    Load disconnection under power 90.4 95.3 92.8 0.912
    Busbar configuration error 90.6 94.5 92.5 0.974
    Unauthorized entry into an energized interval 93.8 94.1 93.9 0.905
    Equipment grounding omission 91.7 93.9 92.8 0.968

Fig. 4.

Fig. 4

Comparison of risk distributions before and after rule embedding. Dashed lines denote the 95th percentile (Inline graphic): Before = 0.65; After = 0.55.

Table 6.

Recognition performance before and after rule constraints.

Model stage Recognized misoperations
(samples)
False positives False negatives Recognized normal operations Accuracy (%)
Before rule embedding 8500 300 1500 3700 87.1
After rule embedding 9200 250 800 3750 92.5

Validation of the rule-driven anti-misoperation model

A fixed-threshold evaluation was adopted to assess time-synchronization adequacy and classification effectiveness. We first analyzed the pre-rule ROC and selected a unified cutoff on that curve (target TPR≈0.90; realized TPRs: pre 0.85, post 0.92). Using this cutoff, we computed confusion matrices for the pre- and post-rule conditions and quantified the changes in FPR, FNR, precision, and recall (in percentage points). To mitigate class-imbalance effects, the confusion matrices were row-normalized (Fig. 5).

Fig. 5.

Fig. 5

Effectiveness of rule-consistency filtering. (a) ROC curves; the solid marker denotes the unified cutoff chosen on the pre-rule ROC, (b) Row-normalized confusion matrices (rows: true labels; columns: predictions); row totals are conserved, (c) Changes in FPR, FNR, precision, and recall (post minus pre) with 95% confidence intervals.

Under this cutoff, rule embedding markedly reduces FPR and improves both precision and recall, with changes remaining controllable and consistent with the engineering objective of prioritizing false-alarm suppression while maintaining recall. The figures validate the marginal effectiveness of the rule-consistency module rather than a system-wide optimum under a globally tuned threshold. Comprehensive system-level results and efficiency comparisons are reported in Table 9.

To ensure consistent timing for comparison, two-stage synchronization (VB followed by UKF) was applied to the validation set. Residual statistics: mean 1.8 ms; standard deviation 0.5 ms; Inline graphic 3.0 ms; maximum < 6 ms. The mean residual is close to zero, indicating no observable system drift. These values serve as engineering-grade timing references; IEEE 1588 (PTP) and GPS-PMU references are used for comparison only and do not constitute a compliance statement or a classification criterion.

Time-synchronization performance

  1. Training Convergence and Stability Analysis

    To assess convergence and stability, three independent training runs were conducted under a fixed configuration, differing only in random seeds; early stopping was triggered by the validation loss. Figure 6 shows the trajectories of training/validation accuracy and loss. The validation set achieved a peak accuracy of 94.69%, and the validation loss was approximately 0.20, indicating no evident overfitting.

  2. Multi-Class Recognition Performance

    Figure 7 presents the row-normalized confusion matrix for nine operational categories on the test set. Class-wise recall ranges from 93.2% to 95.2%, corresponding to a macro-averaged precision of 96.50%, macro-averaged recall of 94.40%, and a macro-averaged F1 of approximately 95%. Minor errors mainly arise from weak-normal samples occasionally predicted as “normal operation” (approximately 1–3% of cases) and mild confusions between insufficient interlock safety and signal delay. Overall, the model distinguishes the nine scenarios with high accuracy, though weak-normal conditions remain improvable.

  3. Comparison across Typical Misoperation Scenarios

    To evaluate discriminative ability under representative misoperation scenarios, per-class Precision, Recall, F1, and AUC are listed in Table 7. TS-MR performs best on forced interlock override and unauthorized grounding disconnection, with AUC of 0.987 and 0.985, and corresponding F1 scores of 93.6% and 94.1%. For unauthorized entry into an energized interval, partial feature overlap with normal operation slightly degrades performance (AUC 0.905, F1 93.9%). Overall, the recognition performance satisfies engineering application requirements.

  4. Correlation between Risk Levels and Misoperation Labels

    To verify the consistency between the risk score computed by Eqs. (1)–(7) and recognition outcomes, we adopt Inline graphic of the normal-state risk-score distribution on the validation set as the reference boundary. Risk thresholds are defined as: high for score ≥ 0.85, medium for 0.65 ≤ score < 0.85, and low for score < 0.65. As shown in Table 8, forced interlock override (0.88) and unauthorized grounding disconnection (0.85) fall into the high-risk category, whereas load disconnection under power scores 0.82 and is classified as medium risk. This mapping aligns with engineering expectations and supports alarm prioritization. Note: these thresholds apply to the task-level risk score and are distinct from the rule-layer engineering thresholds.

Fig. 6.

Fig. 6

Training and validation curves. Blue lines denote step-wise training trajectories; black dots indicate epoch-wise validation aggregation. Vertical axes: accuracy (%) and loss; horizontal axis: training iterations.

Fig. 7.

Fig. 7

Row-normalized confusion matrix on the test set.

Table 8.

Mapping between misoperation labels and risk levels.

Label ID Misoperation type Risk level Reference risk score
1 Unauthorized grounding disconnection High 0.85
2 Forced interlock override High 0.88
3 Unauthorized command execution Medium 0.73
4 Load disconnection under power Medium 0.82
5 Busbar configuration error Medium 0.75
6 Unauthorized entry into an energized interval Medium 0.70

Recognition performance and explainability

To improve diagnostic efficiency, an XAI visualization module is integrated into the inference stage to output three types of information: (1) attention heatmaps highlighting Top-K evidence; (2) modality contribution quantified by permutation importance; and (3) a comparison of the risk-probability curve with the preset decision threshold, as shown in Fig. 8.

Fig. 8.

Fig. 8

XAI-based explainability analysis. (a) Attention with Top-K evidence, (b) Modality contribution (permutation importance), (c) Evidence timeline: risk probability vs threshold; Top-K hit = 0.40; TLE = N/A.

In the illustrated case, the modality-level contribution weights are: SCADA 0.56, PMU 0.08, Video 0.10, Voice 0.19, and Ticket 0.07. The Top-K hit rate is 0.40 (i.e., 2 of 5 evidence points match the ground-truth labels). The risk-probability curve does not exceed the threshold anywhere within the window; therefore, the event is judged as not triggering a misoperation, and TLE is recorded as N/A.

The explainability outputs are consistent with the operations log: the decision relies primarily on SCADA, with voice and operation tickets providing auxiliary cues, while PMU contributes weakly in this scenario. Overall, the XAI results align with the model’s predictions and help operations and maintenance (O&M) personnel rapidly trace evidence and interpret decisions.

Risk-response latency analysis

To evaluate the temporal responsiveness of TS-MR, the risk curve was computed according to Eqs. (1)–(7), with thresholds set to Inline graphic and Inline graphic. The first threshold-crossing moment was denoted Inline graphic, and the actual lockout trigger moment Inline graphic; the response delay was defined as Inline graphic. Figure 9 shows that the curve remains below the threshold in the low-risk scenario; sparse events occur in the medium-risk scenario; and in the high-risk scenario, lockouts are frequent while the system maintains stable tracking. Cross-scenario statistics indicate that Inline graphic mainly falls within 1.5–3.0 s, with Inline graphic, Inline graphic, and Inline graphic These results demonstrate stable and predictable temporal response across different risk intensities, meeting engineering application requirements.

Fig. 9.

Fig. 9

Risk-response latency analysis. (a) Risk curve and threshold under the low-risk scenario; (b) misoperation events under the medium-risk scenario (red triangles mark event timestamps); (c) lockout-state sequence under the high-risk scenario (blue vertical lines indicate lockout triggers).

Independent HIL cross-validation

To verify deployability, we conducted an independent check in a hardware-in-the-loop (HIL) environment focusing on interface connectivity and temporal consistency; these results are not included in the main evaluation. Figure 10(a) shows the HIL risk timeline aligned to the offline time base without applying a decision threshold. Figure 10(b) compares ROC shapes of offline and HIL runs; the AUC values (0.540 offline, 0.516 HIL) are reported only as a connectivity sanity check and are not used as classification performance. Figure 10(c) reports end-to-end latency with batch size 1: the offline distribution mainly lies within 60–100 ms (mean about 80 ms); HIL shifts slightly to the right (median about 90 ms) with a few samples exceeding 120 ms. Overall, interfaces and timing remain stable in HIL, supporting engineering-grade deployability.

Fig. 10.

Fig. 10

Independent HIL cross-validation. (a) HIL risk timeline aligned to the offline time base, (b) ROC curves, offline versus HIL; AUCs are shown only to verify connectivity and label alignment, not as performance metrics, (c) End-to-end latency (batch size 1).

Comprehensive performance comparison under the unified evaluation protocol

To ensure comparability, all models adopt identical preprocessing, time alignment and window configurations, and a unified protocol for training, validation, and testing. Random seeds are fixed; class imbalance is handled by weighted cross-entropy; and early stopping is applied to avoid overfitting. Three baselines (CNN-BiLSTM, ConvLSTM, GAT) are tuned within the same search space: learning rate Inline graphic, network depth Inline graphic, dropout Inline graphic, and hidden dimension Inline graphic. The best configurations are used for testing to avoid bias from “insufficient baseline tuning.” Under this unified protocol, TS-MR attains accuracy 94.69%, F1-score 93.8%, and AUC 0.977, outperforming all baselines. End-to-end latency (batch = 1) is about 80 ms, the core forward pass is about 42 ms, and computational complexity is 3.4 GFLOPs, achieving a strong accuracy–efficiency trade-off. Compared with the three baselines, accuracy improves by 0.9–3.7 percentage points and AUC increases by 0.024–0.057, consistent with previous statistical checks.

Conclusion

To address temporal misalignment in multi-source heterogeneous data, complex operational logic, and the risk of misoperations during microgrid switching, this study proposes a time-synchronized misoperation recognition framework (TS-MR). The framework integrates procedural and interlocking constraints with dual-stage temporal alignment and multimodal fusion, forming an end-to-end pipeline comprising rule-based screening, temporal alignment, feature discrimination, and constraint optimization. This enables reliable misoperation detection and quantitative risk assessment suitable for data-intensive, disturbance-prone edge-intelligent environments.

  1. Temporal alignment and feature modeling: A dual-stage synchronization mechanism combining variational Bayesian inference and UKF reduces cross-modal temporal bias and enhances sensitivity to key triggers. On this basis, the Transformer aligns multimodal semantics, the TCN models sequential dynamics, and the GNN captures equipment topology and interlocking dependencies, yielding stable performance in high-risk scenarios such as forced interlock override and unauthorized entry into energized intervals. Temporal alignment robustness is benchmarked against IEEE 1588 PTP and PMU timing referenced to GPS-disciplined UTC for magnitude comparison and robustness characterization.

  2. Rule screening and risk priors: The rule module performs deterministic compliance checks to ensure that operation sequences satisfy procedural and interlocking constraints. Monte Carlo-based risk assessment is used only as an independent probabilistic prior during training for sample weighting and priority scheduling; it does not participate in rule judgments or real-time compliance processes. Together with consistency checks and perturbation analysis—supported by attention heatmaps, permutation importance, and node-contribution evidence—the model provides auditable support for suppressing false alarms and improving robustness.

  3. Comprehensive performance and efficiency: Under a unified evaluation protocol, TS-MR outperforms comparative models in both accuracy and efficiency. End-to-end latency meets engineering real-time requirements, and computational complexity is constrained and reported using GFLOPs. The engineering standards and link-budget references are used solely for quantitative comparison and benchmarking, and do not constitute a compliance statement.

  4. Limitations and outlook: This study primarily uses real and laboratory measurements; HIL data are reserved for independent cross-validation, and public datasets are used only for cross-domain robustness calibration, without involvement in training, validation, or testing. Synthetic samples are employed in the training set solely for imbalance handling. Although StarSim-HIL testbed validation has been completed, larger-scale field verification with extended PMU and SCADA datasets is still required to assess long-term generalization and engineering feasibility. Future work will target multi-station coordination, operator behavior prediction, and strategy co-optimization, incorporating additional HIL and field data within a cloud-edge collaborative architecture to enhance transferability and real-time deployability.

In summary, TS-MR achieves a balanced trade-off among accuracy, latency, and interpretability, providing a reusable technical pathway for trustworthy intelligent operation in microgrid switching scenarios.

Supplementary Information

Below is the link to the electronic supplementary material.

Supplementary Material 1 (284.9KB, docx)

Author contributions

Hongzhao Yang was responsible for the overall research framework design and manuscript writing. Zhan Zhang handled data processing and model simulation. Shijie Zhang contributed to the development and analysis of the time synchronization algorithm. Rui Liang was in charge of the anti-misoperation rule modeling and experimental validation. All authors have read and approved the final version of the manuscript.

Funding

This work was supported by Hunan Provincial Natural Science Foundation of China (No. 2024JJ8025) and the National Natural Science Foundation of China (Nos. 7217010719, 72171026).

Data availability

The datasets generated and/or analyzed during the current study are available from the corresponding author on reasonable request.

Declarations

Competing interests

The authors declare no competing interests.

Footnotes

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Das, S. S. et al. A comparative analysis of the efficient coordination of renewable energy and electric vehicles in a deregulated smart power system. Energy Rep.13, 3136–3164. 10.1016/j.egyr.2025.02.047 (2025). [Google Scholar]
  • 2.Adeyinka, A. M. et al. Advancements in hybrid energy storage systems for enhancing renewable energy-to-grid integration. Sustain. Energy Res.11, 26. 10.1186/s40807-024-00120-4 (2024). [Google Scholar]
  • 3.Khalid, M. Smart grids and renewable energy systems: Perspectives and grid integration challenges. Energy Strateg. Rev.51, 101299. 10.1016/j.esr.2024.101299 (2024). [Google Scholar]
  • 4.Uddin, M. et al. Microgrids: A review, outstanding issues and future trends. Energ. Strat. Rev.49, 101127. 10.1016/j.esr.2023.101127 (2023). [Google Scholar]
  • 5.Choudhury, S., Dash, T., Sahu, P. K. & Sahoo, P. K. A comprehensive review on issues, investigations, control and protection trends, technical challenges and future directions for Microgrid technology. Int. Trans. Elect. Energy Syst.30(9), e12446. 10.1002/2050-7038.12446 (2020). [Google Scholar]
  • 6.Abbasi, M. et al. Optimizing microgrid performance: a multi-objective strategy for renewable energy integration and cost reduction in Khyber Pakhtunkhwa. Sci. Rep.15, 11234. 10.1038/s41598-025-00118-y (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Rangu, S. K., Lolla, P. R., Dhenuvakonda, K. R. & Singh, A. R. Recent trends in power management strategies for optimal operation of distributed energy resources in microgrids: A comprehensive review. Int. J. Energy Res.44(13), 9889–9911. 10.1002/er.5649 (2020). [Google Scholar]
  • 8.She, F. Li, H. Cui, J. Zhang and R. Bo, "Fusion of Microgrid Control With Model-Free Reinforcement Learning: Review and Vision," in IEEE Transactions on Smart Grid, vol. 14, no. 4, pp. 3232–3245, July 2023, 10.1109/TSG.2022.3222323.
  • 9.Li, C., Rakhra, P., Norman, P. J., Burt, G. M. & Clarkson, P. Multi-Sample Differential Protection Scheme in DC Microgrids. IEEE J. Emerg. Select. Topics in Power Electron.9(3), 2560–2573. 10.1109/JESTPE.2020.3005588 (2021). [Google Scholar]
  • 10.Somanna, B. et al. Optimized fault detection and control for enhanced reliability and efficiency in DC microgrids. Sci Rep15, 31161. 10.1038/s41598-025-98006-y (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Ćetenović, D., Zhao, J., Levi, V., Liu, Y. & Terzija, V. Variational Bayesian Unscented Kalman Filter for Active Distribution System State Estimation. IEEE Trans. Power Syst.40(1), 476–491. 10.1109/TPWRS.2024.3406399 (2025). [Google Scholar]
  • 12.A. Vafamand, B. Moshiri and N. Vafamand, "Fusing Unscented Kalman Filter to Detect and Isolate Sensor Faults in DC Microgrids With CPLs," in IEEE Transactions on Instrumentation and Measurement, vol. 71, pp. 1–8, 2022, Art no. 3503608, 10.1109/TIM.2021.3135547.
  • 13.G. K. Michelon, W. K. G. Assunção, P. Grünbacher and A. Egyed, "Analysis and Propagation of Feature Revisions in Preprocessor-based Software Product Lines," 2023 IEEE International Conference on Software Analysis, Evolution and Reengineering (SANER), Taipa, Macao, 2023, pp. 284–295, 10.1109/SANER56733.2023.00035.
  • 14.Wang, Y. et al. Multimodal anomaly detection in complex environments using video and audio fusion. Sci Rep15, 16291. 10.1038/s41598-025-01146-4 (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Wen, X. et al. Deep learning-driven hybrid model for short-term load forecasting and smart grid information management. Sci Rep14, 13720. 10.1038/s41598-024-63262-x (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Talebi, S., & Zhou, K. (2025). Graph Neural Networks for Efficient AC Power Flow Prediction in Power Grids. arXiv preprint arXiv:2502.05702. https://arxiv.org/abs/2502.05702.
  • 17.Bhagwat, A. et al. An online 11 kv distribution system insulator defect detection approach with modified YOLOv11 and mobileNetV3. Sci Rep15, 15691. 10.1038/s41598-025-99756-5 (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Bantwal, A. S., Bhayadia, A. K. & Meng, H. Importance of considering temporal variations in pulse wave velocity for accurate blood pressure prediction. Ann. Biomed. Eng.53, 1080–1094. 10.1007/s10439-025-03681-7 (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Yang, H., Zhang, Z., Liang, R. & Zhao, W. Research on time-of-use compensation pricing strategies for load aggregators based on user demand response. Front. Energy Res.12, 1442194. 10.3389/fenrg.2024.1442194 (2024). [Google Scholar]
  • 20.Lv, Y., Li, K., Zhao, H. & Lei, H. A multi-stage constraint-handling multi-objective optimization method for resilient microgrid energy management. Appl. Sci.14, 3253. 10.3390/app14083253 (2024). [Google Scholar]
  • 21.Cheng, G., Lin, Y., Abur, A., Gómez-Expósito, A. & Wu, W. A survey of power system state estimation using multiple data sources: PMUs, SCADA, AMI, and Beyond. IEEE Trans. Smart Grid15(1), 1129–1151. 10.1109/TSG.2023.3286401 (2024). [Google Scholar]
  • 22.El-Agamy, R. F. et al. Comprehensive analysis of digital twins in smart cities: a 4200-paper bibliometric study. Artif. Intell. Rev.57, 154. 10.1007/s10462-024-10781-8 (2024). [Google Scholar]
  • 23.Joshi, A., Capezza, S., Alhaji, A. & Chow, M.-Y. Survey on AI and machine learning techniques for microgrid energy management systems. IEEE/CAA J. Automatica Sinica10(7), 1513–1529. 10.1109/JAS.2023.123657 (2023). [Google Scholar]
  • 24.Manoj, R. et al. Measurement of pressure dependent variations in local pulse wave velocity within a cardiac cycle from forward travelling pulse waves. Sci Rep15, 3066. 10.1038/s41598-025-87143-z (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.X. Wang, H. Zhu, X. Luo and X. Guan, “Data-Driven-Based Detection and Localization Framework Against False Data Injection Attacks in DC Microgrids, ” in IEEE Internet of Things Journal, vol. 12, no. 17, pp. 36079–36093, 1 Sept.1, 2025, 10.1109/JIOT.2025.3579915.
  • 26."IEEE Standard for the Specification of Microgrid Controllers," In IEEE Std 2030.7–2017 , vol., no., pp.1–43, 23 April 2018, 10.1109/IEEESTD.2018.8340204.
  • 27."IEEE Standard for Interconnection and Interoperability of Distributed Energy Resources with Associated Electric Power Systems Interfaces," In IEEE Std 1547–2018 (Revision of IEEE Std 1547–2003) , vol., no., pp.1–138, 6 April 2018, 10.1109/IEEESTD.2018.8332112.
  • 28.Sun, P., Zhao, B. & Li, X. Research on multi-source heterogeneous data fusion method of substation based on cloud edge collaboration and AI technology. Discov Appl Sci7, 262. 10.1007/s42452-025-06725-8 (2025). [Google Scholar]
  • 29.IEEE Standards Association. IEEE Standard for a Precision Clock Synchronization Protocol for Networked Measurement and Control Systems. IEEE Std 1588–2019 (2019). 10.1109/IEEESTD.2020.9120376
  • 30.IEC & IEEE. Communication networks and systems for power utility automation. Part 9–3: Precision time protocol profile for power utility automation. IEC/IEEE 61850–9–3:2016 (2016). 10.1109/IEEESTD.2016.7479438
  • 31.IEC & IEEE. Measuring relays and protection equipment. Part 118–1: Synchrophasor for power systems. Measurements. IEC/IEEE 60255–118–1:2018 (2018). 10.1109/IEEESTD.2018.8577045
  • 32.A. Scandelli, C. Lombardi, P. Castiglioni, G. Merati, F. Villa and A. Faini, "Comparison of Methods to Estimate Arterial Blood Pressure From the Pulse Arrival Time by Finger Photoplethysmography," in IEEE Sensors Journal, vol. 25, no. 1, pp. 783–791, 1 Jan.1, 2025, 10.1109/JSEN.2024.3486565.
  • 33.PTPd Development Team. PTPd: Precision Time Protocol daemon, Version 2.3.1. GitHub repository, https://github.com/ptpd/ptpd (2023). Accessed 20 June 2025.
  • 34.Soleymannejad, M.; Sadrian Zadeh, D.; Moshiri, B.; Sadjadi, E.N.; Herrero, J.G.; López, J.M.M. State Estimation Fusion for Linear Microgrids over an Unreliable Network. Energies 2022, 15, 2288. 10.3390/en15062288.
  • 35.Rozzi, E., Minuto, F. D. & Lanzini, A. Techno-economic dataset for hydrogen storage-based microgrids. Data Brief56, 110795. 10.1016/j.dib.2024.110795 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Bashir, A. et al. Power, voltage, frequency and temperature dataset from Mesa Del Sol microgrid. Dryad 10.5061/dryad.fqz612jzb (2023).

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Citations

  1. Bashir, A. et al. Power, voltage, frequency and temperature dataset from Mesa Del Sol microgrid. Dryad 10.5061/dryad.fqz612jzb (2023).

Supplementary Materials

Supplementary Material 1 (284.9KB, docx)

Data Availability Statement

The datasets generated and/or analyzed during the current study are available from the corresponding author on reasonable request.


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES