Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2025 Feb 5;15:4337. doi: 10.1038/s41598-025-88450-1

Artificial intelligence-driven ensemble deep learning models for smart monitoring of indoor activities in IoT environment for people with disabilities

Munya A Arasi 1,, Hussah Nasser AlEisa 2, Amani A Alneil 3,4, Radwa Marzouk 5
PMCID: PMC11799421  PMID: 39910242

Abstract

Disabled persons demanding healthcare is a developing global occurrence. The support in longer-term care includes nursing, intricate medical, recovery, and social help services. The price is large, but advanced technologies can aid in decreasing expenditure by certifying effective health services and enhancing the superiority of life. The transformative latent of the Internet of Things (IoT) prolongs the existence of nearly one billion persons worldwide with disabilities. By incorporating smart devices and technologies, the IoT provides advanced solutions to tackle numerous tasks challenged by individuals with disabilities and promote equality. Human activity detection methods are the technical area which studies the classification of actions or movements an individual achieves over the recognition of signals directed by smartphones or wearable sensors or over images or video frames. They are efficient in certifying functions of detection of actions, observing crucial functions, and tracking. Conventional machine learning and deep learning approaches effectively detect human activity. This study develops and designs a metaheuristic optimization-driven ensemble model for smart monitoring of indoor activities for disabled persons (MOEM-SMIADP) model. The proposed MOEM-SMIADP model concentrates on detecting and classifying indoor activities using IoT applications for physically challenged people. First, data preprocessing is performed using min–max normalization to convert input data into useful format. Furthermore, the marine predator algorithm is employed in feature selection. For the detection of indoor activities, the proposed MOEM-SMIADP model utilizes an ensemble of three classifiers, namely the graph convolutional network model, long short-term memory sequence-to-sequence (LSTM-seq2seq) method, and convolutional autoencoder. Eventually, the hyperparameter tuning is accomplished by an improved coati optimization algorithm to enhance the classification outcomes of ensemble models. A wide range of experiments was accompanied to endorse the performance of the MOEM-SMIADP technique. The performance validation of the MOEM-SMIADP technique portrayed a superior accracy value of 99.07% over existing methods.

Keywords: Indoor activity detection, Disabled persons, Ensemble models, Improved coati optimization algorithm, Internet of Things

Subject terms: Computational platforms and environments, Machine learning

Introduction

Globally, more than one billion individuals, nearly 15% of the population, live with more than one disability, as per the World Health Organization (WHO). These diseases can exhibit initially in childhood or improve in older age, like the function of an impaired hand due to stroke1. Day by day, people with disabilities substantially struggle to manage their home appliances. Thus, classical houses were transformed into smart homes to enhance living standards for people with disabilities. In past years, the purpose of IoT technology has been to permit the interaction between gadgets without the necessity for human involvement2. Nowadays, IoT technology is incorporated with home gadgets to allow these devices to control the internet remotely. IoT defines physical object networks or things with software, sensors, and other technology developed to share and connect data with other devices or systems through the internet3. These gadgets comprise light switches that respond to thermostats or turn-off and on commands that can modify the indoor temperature to decrease energy consumption. Diverse authors represent multiple solutions for assisting disabled persons in regulating gadgets remotely through IoT depending on the consumer voice or the smartphone Graphical User Interface (GUI)4. Human Activity Recognition (HAR) is the domain of science that examines action or movement identifications undertaken to detect signals sent by smartphones or wearable sensors through image or video frames5.

These actions are implemented indoors, like sitting, walking, standing, and stairs. Also essential to identify where the practical activities are implemented. Computer technology and human movements were utilized to understand artificial vision. HAR has many applications like anti-terrorist security, surveillance, assistance, and lifelogging6. These methods have been verified as helpful for offering effective home care for disabled persons and indoor tracking methods. The number of disabled persons and older adults increases, thus defining a necessity to support those people who are losing autonomy and want to stay alive in their houses, requiring continuous assistance in the real world7. Compared with indoor and outdoor localization, it is difficult for indoor communication channels to diverge considerably from their surroundings. It is based on several factors, like construction materials, building structure, and room layout. The Indoor Positioning System (IPS) presents real-world localization of people or objects, and spaces are surrounded by diverse environments, utilizing a network for receivers and transmitters8. HAR can be defined as the art of identifying and naming activities using artificial intelligence (AI) from the activities collected from raw data by employing several resources. ML and DL progress have made meaningful task feature extractors from the raw sensor data. The increasing global prevalence of disabilities highlights the requirement for effective solutions that improve the quality of life for affected individuals. People with physical or cognitive impairments often encounter challenges in managing everyday tasks, comprising controlling their home environment9. With the rise of connected devices and smart technologies, there is a growing opportunity to develop systems that offer autonomy and support to people with disabilities. Intelligent monitoring systems can facilitate enhanced home management, ensuring safety and independence. Integrating advanced DL methods within an IoT ecosystem can revolutionize how smart homes are designed, providing tailored solutions for diverse needs10.

This study develops and designs a Metaheuristic Optimization-Driven Ensemble Model for Smart Monitoring of Indoor Activities for Disabled Persons (MOEM-SMIADP) model. The proposed MOEM-SMIADP model concentrates on detecting and classifying indoor activities using IoT applications for physically challenged people. First, data preprocessing is performed using min–max normalization to convert input data into useful format. Furthermore, the marine predator algorithm (MPA) is employed in feature selection. For the detection of indoor activities, the proposed MOEM-SMIADP model utilizes an ensemble of three classifiers, namely the graph convolutional network (GCN) model, long short-term memory sequence-to-sequence (LSTM-seq2seq) method, and convolutional autoencoder (CAE). Eventually, the hyperparameter tuning is accomplished by an improved coati optimization algorithm (ICOA) to enhance the classification outcomes of ensemble models. A wide range of experiments was accompanied to endorse the performance of the MOEM-SMIADP technique. The key contribution of the MOEM-SMIADP technique is listed below.

  • The MOEM-SMIADP model utilizes min–max normalization to preprocess data by scaling it to a specific range, ensuring balanced feature contributions during learning. This approach improves subsequent algorithms’ performance and convergence speed, specifically in handling diverse feature scales effectively. It plays a significant role in improving the accuracy and stability of the ensemble classifiers.

  • The MOEM-SMIADP approach employs the MPA method to choose the most relevant features, effectively mitigating data dimensionality and eliminating noise or irrelevant information. This optimization-driven approach improves computational efficiency and ensures enhanced performance of subsequent classifiers. It significantly improves the model’s capability to focus on critical features, enhancing overall detection accuracy.

  • The MOEM-SMIADP method combines three classifiers to utilize diverse feature extraction capabilities. The GCN captures spatial and structural relationships, the LSTM-seq2seq handles temporal sequences and long-term dependencies in time-series data, and the CAE extracts latent features for complex pattern detection. This integration improves the accuracy and robustness of the model in indoor activity recognition.

  • The MOEM-SMIADP model utilizes an ICOA model for hyperparameter tuning, ensuring optimal configurations for the ensemble classifiers. This approach improves the model’s accuracy, performance, and efficiency by finding the optimal parameter settings. It plays a key role in improving the detection capability and robustness of the indoor activity recognition system.

  • The MOEM-SMIADP methodology outperforms by integrating diverse methods such as MPA-based feature selection, ICOA-based hyperparameter tuning, and an ensemble of three state-of-the-art classifiers, namely GCN, LSTM-seq2seq, CAE, to address indoor activity detection. This synergistic approach uniquely incorporates spatial, temporal, and latent feature extraction techniques with advanced optimization algorithms, resulting in significant improvements in detection accuracy and efficiency compared to traditional methods.

Related works

In11, a strong DL structure known as a Multiple Spectrogram Fusion Network (MSF-Net) is presented for fine activity recognition and coarse utilizing Channel State Information (CSI). Initially, a dual-stream framework integrating DWT and short-time Fourier transform is introduced to highlight abnormal information in the CSI data. Formerly, a Transformer was applied as the backbone to remove higher-level features efficiently. Berkani et al.12 developed an intelligent method to monitor air quality and classify activities in indoor surroundings employing the DL method depending on a 1D Convolutional Neural Network (1D-CNN). This method incorporates six sensors to collect measurement parameters that are eventually trained in a 1D CNN method for activity detection. This projected method boasts an edge-deployable and lightweight design, making it standard for real-world applications. Sun and Chen13 developed a novel asynchronous detection approach, the Rapid Response Elderly Safety Monitoring (RESAM) technique. During the primary analysis of inertial sensor data utilizing multi-class classifiers and Kernel Principal Component Analysis (KPCA), this method effectively decreases the processing period and lowers the FNR. Then, decision-level data fusion was performed, integrating skeleton image investigation depending on the primary stage’s inertial sensor data and ResNet. In14, data processing approaches suited for a non-invasive indoor noisy sound examination method operating edge environments were developed. To accomplish this, MFCC and Mel-spectrogram-based methods for classifying sound environments are applied to compare their performance depending on optimizations and diverse preprocessing parameters. In15, a BC and IoT-based Assisted Living System (BIoT-ALS) is presented utilizing 6G communication. These nodes in the projected model use smart contracts to particularize norms of interaction while working together to offer computing resources and storage. Kan et al.16 developed an innovative method employing dual Kinect V2, developed by progressed ensemble learning models and advanced Transmission Control Protocol (TCP). Data-adaptive adjustment mechanism, embedded in localization results, to decrease self-occlusion in dynamic orientations and amalgamation of the RF and bat models, offering novel action detection approaches for complex scenarios. Srinivasan et al.17 developed an innovative approach to enhance outdoor comfort by associating adaptive thermal apparel with RFR and IoT. Wearing inflexible classical outdoor gear might be a real pain for those living in diverse forms of the world. This method introduces an intelligent garment method, which can change its thermal insulation in the real world by employing data collected from IoT sensors. An RFR method was used to determine the useful thermal settings for the garments, depending on collected data, which included environmental aspects like temperature, wind speed, and humidity.

Manimaran et al.18 aimed to support the elderly by tracking their activities in both outdoor and indoor surroundings. A semi-supervised DL structure was introduced for better HAR outcomes, which efficiently uses the incorrectly labelled sensor information and fine-tunes the classifier learning method. Xiao et al.19 present a framework for activity recognition and health monitoring using smartphone accelerometer data, utilizing BiLSTM and Bayesian optimization to improve the performance and fine-tune the model. Shereef, Varghese, and Kamalraj20 explore the role of IoT and cloud computing in healthcare, concentrating on their application in diagnosing sleep apnea while addressing benefits, challenges, and their potential to revolutionize sleep medicine. Rezaee et al.21 present an optimized BiLSTM model with Grey Wolf Optimizer (GWO) for real-time student activity classification and health monitoring using accelerometer data, validated on UCI-HAR and WISDM datasets. Anitha et al.22 enhance fall detection in the elderly by incorporating multiple sensors with AI and ML methods for real-time monitoring, accuracy, and adaptability while exploring future sensor integration for improved health tracking. Maddeh et al.23 propose an ensemble DL model to detect a patient’s mobility state using sensors in a smartbed, distinguishing between sleeping, standing, sitting, walking, and emergency states for improved accuracy. Akhmetshin et al.24 present EADL-FDC, using DL and evolutionary algorithms for fall detection, with SPA-Net for feature extraction, SOS for parameter selection, DBN for classification, and MFO for hyperparameter tuning. Namoun et al.25 propose an ensemble meta-learning model to select the best IoT services for disabled students in education, considering their unique needs. Jawad et al.26 developed a sustainable greenhouse model that optimizes energy consumption while ensuring optimal plant growth conditions using the Artificial Bee Colony (ABC) optimization technique and a fuzzy controller to regulate environmental factors. Kao et al.27 explore drowning prevention technology using embedded systems, AI, and IoT for real-time monitoring and alerting. Computer vision and DL improve image recognition for identifying drowning situations, while IoT connectivity improves system intelligence and rescue efficiency. Yazici et al.28 present an e-health framework utilizing IoT-based inertial, ECG, and video sensors for real-time monitoring of elderly and disabled individuals, employing edge computing for efficient data analysis and ensuring privacy by activating sensors only when needed.

The reviewed studies highlight significant IoT, AI, and DL improvements for healthcare, activity recognition, and assistive systems. However, limitations persist, comprising challenges in real-time processing, scalability, and adaptability to diverse environments and user needs. Many solutions face privacy concerns, high computational costs, and energy inefficiency, specifically in resource-constrained settings. Furthermore, integrating multi-sensor data and optimizing models for dynamic and heterogeneous conditions remain underexplored. Research gaps encompass the requirement for robust frameworks that ensure privacy, energy efficiency, and adaptability while giving seamless integration of advanced technologies like 6G, blockchain, and edge computing for scalable, real-time applications across broader use cases.

The proposed method

This study develops a MOEM-SMIADP model. The proposed model concentrates on detecting and classifying indoor activities using IoT applications for physically challenged people. It encompasses four steps: data normalization, MPA-based feature selection, an ensemble of classification models, and parameter selection using ICOA. Figure 1 illustrates the workflow of the MOEM-SMIADP model.

Fig. 1.

Fig. 1

Workflow of MOEM-SMIADP model.

Min–max normalization

At first, the data preprocessing executes min–max normalization to convert input data into useful format29. This is chosen because it can scale data to a fixed range, usually [0, 1] or [− 1, 1], ensuring uniform contribution from all features during the learning process. This technique is particularly effectual when features have varying scales, preventing dominant features from overshadowing smaller ones. Unlike standardization, which transforms data based on mean and standard deviation, min–max normalization preserves the data’s distribution and relationships, making it ideal for algorithms sensitive to an absolute scale, such as GCNs and LSTMs. Additionally, it improves the convergence speed of optimization algorithms, mitigating training time and improving stability. Its computational efficiency makes it well-suited for large datasets and real-time applications.

Min–max normalization is an effective data preprocessing mode, which measures feature values to a definite range from [0, 1], maintaining the relations within the data. In smart observing of indoor actions utilizing IoT applications, this model certifies consistency across data gathered from numerous sensors, which might have dissimilar ranges or units. For disabled persons, accurate recognition of anomalies and activities is vital, and min–max normalization decreases the impact of outliers and noise, thereby improving the excellence of data. It enhances the performance of ML techniques by delivering standardized inputs, permitting methods to classify refined variations in activity patterns. This ensures adaptive, reliable, and effective monitoring methods, which provide individual requirements, promoting safety and independence.

Feature selection using MPA

Furthermore, the MPA performs the feature selection process30. This technique is chosen because it can effectually identify the most relevant features while discarding noisy or redundant ones. MPA replicates the intelligent foraging strategies of marine predators, effectively balancing exploration and exploitation during optimization. Unlike conventional techniques like correlation-based or wrapper methods, MPA can handle high-dimensional data and intrinsic interactions between features. This mitigates computational overhead and enhances the model’s performance by focusing only on significant features. Its adaptability and robustness in diverse data scenarios make it particularly appropriate for improving classifier accuracy and ensuring improved generalization.

The marine predator’s foraging movements stimulate MPA. Predators often switch between dual motion patterns: Brownian motion (BM), which involves consecutive moves in a similar position that improves exploitation, and Levy motion (LM), which involves short motions followed by higher jumps that increase exploration.

Stage (1). Initialization: the search space is packed using the randomly and uniformly distributed primary solutions.

Stage (2). The prey matrices are upgraded in stage (1), considered by higher-velocity ratios. This upgrade occurs in the first third of iterations while exploring problems.

graphic file with name M1.gif 1
graphic file with name M2.gif 2

Meanwhile, RB denotes the vector of random numbers according to the standard distribution of BM, Inline graphic, and Inline graphic signifies a vector of uniform randomly formed integers amongst 0 and 1.

Stage (3). Stage (2) upgrade is identified as the transitional optimizer stage, whereas the model shifts from exploration to exploitation. This procedure occurs in the second third of the iterations.

graphic file with name M5.gif 3
graphic file with name M6.gif 4

Inline graphic, an arbitrary value vector according to the LM standard distribution, is multiplied through the prey during the step equation. For the second half of populations, it can be upgraded utilizing Eqs. (5) and (6).

graphic file with name M8.gif 5
graphic file with name M9.gif 6

Determine that RB and the matrix of Elite in Eq. (5) are multiplied to mimic the BM. CF denotes the parameter for controlling the step dimensions and can be upgraded with Eq. (7).

graphic file with name M10.gif 7

Stage (4). Stage 3, the last third of iterations, is measured as the last phase of the optimizer procedure. The predator moves to utilize LM, and subsequently, the prey matrix can be upgraded using Eqs. (8) and (9).

graphic file with name M11.gif 8
graphic file with name M12.gif 9

The Inline graphic and Inline graphic matrix multiplication mimics the predator’s motion in Lévy’s tactic.

Stage (5). Finalization Process: The optimal solutions are continually included in the matrix of the Elite succeeding all iterations. After achieving the maximal iteration counts, the last solution with the top fitness function (FF) will become apparent.

The FF imitates the accuracy of the classifier and the sum of the chosen features. It exploits the classifier’s accuracy and diminishes the set dimensions of the chosen features. Consequently, the FF is used to assess individual solutions, which is given in Eq. (10).

graphic file with name M15.gif 10

Here, Inline graphic refers to the classification rate of error using the chosen features. Inline graphic has been computed as the ratio of improper classifications to the number of classifications between 0 and 1. Inline graphic refers to the quantity of chosen features, and Inline graphic means the total amount of features. Inline graphic is applied for controlling the import of classifier excellence and sub-set length.

Indoor activities detection using ensemble models

For the detection of indoor activities, the proposed MOEM-SMIADP model applies an ensemble of three classifiers: the GCN model, the LSTM-seq2seq method, and the CAE classifier. This ensemble technique is chosen for its ability to capture diverse and complementary data features. GCN outperforms in learning spatial and structural patterns, LSTM-seq2seq effectually handles temporal sequences and long-term dependencies, and CAE identifies complex latent features. This integration outperforms single-model approaches by utilizing the merits of each classifier, improving detection accuracy and robustness. Unlike standalone methods, the ensemble approach mitigates the risk of overfitting and adapts better to complex indoor activity patterns. Its versatility makes it specifically effective in handling multimodal and dynamic datasets.

GCN model

GCN is based on CNN, which concentrates on the convolutional process of operation31. It applies functional mapping to make novel node information by combining the neighbouring and the present node data. The space-based GCN straight aggregates the handling of graph-structured data based on nodes and edges, significantly decreasing the calculation sum, and is currently frequently utilized in networks. Here, a GCN‐based model is presented for learning to carry out the generator’s positions by removing power grid graph data. Applying the topologic architecture and bus attributes of the power grid, a graphical representation of the power grid Inline graphic can be originated, Inline graphic denotes a collection of buses in Inline graphic Inline graphic represents a collection of bus feature sizes. Bus within the power grid is separated into dual types according to the absence or presence of the generator. This attribute is about loaded data and generator data related to limitations. The message concerning transmission lines related to limits also needs to be considered. Hence, the capability of transmission lines among buses must be combined into the feature matrix. The feature matrix construction Inline graphic for GCN can be established as shown.

graphic file with name M26.gif 11
graphic file with name M27.gif 12

When there are no lines of transmission among nodes Inline graphic and Inline graphic is equivalent to Inline graphic Inline graphic denotes binary variable to designate if a generator lies in bus Inline graphic; when a generator occurs at node Inline graphic is equivalent to Inline graphic, or else it becomes Inline graphic

Owing to the distinctive measurement elements related to the various features within the input feature data, straight calculations and evaluations are not possible. To tackle this problem, a normalization method is used to preprocess Inline graphic. Essentially, this process is not utilized for the complete dataset. Still, it is performed distinctly for every feature. This method can improve the efficiency of the training. The succeeding equation is applied for the normalization procedure.

graphic file with name M37.gif 13

The GCN input is Inline graphic, and the initial hidden feature vector Inline graphic is gained after the initial graph convolution layer (GCL). This method has been executed by aggregating the neighbouring bus features and then passing over a linear transformation. Afterward numerous convolution layers, the last output outcome is gained over the fully connected layer (FCL). Meanwhile, Inline graphic is not normalized; handling Inline graphic would alter the feature vector scales. To resolve these issues, A and the matrix of degree Inline graphic are included and then standardized over the node matrix of degree Inline graphic

graphic file with name M44.gif 14

Formerly, the computation of every GCL is stated as follows:

graphic file with name M45.gif 15

While Inline graphic denotes the activation function. Figure 2 illustrates the infrastructure of the GCN model.

Fig. 2.

Fig. 2

Structure of GCN.

As the preprocessed Inline graphic aids as the input to GCN, each bus’s novel Inline graphic dimensional feature vectors are gained over the transformation of GCL and FCL, and the attainment of the forecast promise is characterized as shown.

graphic file with name M49.gif 16

While Inline graphic specifies the GCN process, Inline graphic specifies the removal of consistent likelihood vectors.

GCN is applied to discover the binary variable patterns, so the learning task corresponds to a multi-class binary classification task.

graphic file with name M52.gif 17

Inline graphic represents the variance between the target and the predicted values.

LSTM-seq2seq model

The LSTM-seq2seq framework contains a decoder and encoder32. In encoding NN, the input sequence Inline graphic with the time step counts Inline graphic is read a single time step at a period. Finally, the hidden layer Inline graphic creates a higher dimension Inline graphic vector, encoded to signify the data from the input series. The decoder NN framework proceeds the vector Inline graphic as the input to attain the resultant series Inline graphic across the loop directed. This computation, which involves the LSTM-seq2seq framework, is followed,

graphic file with name M60.gif 18
graphic file with name M61.gif 19
graphic file with name M62.gif 20

While Inline graphic and Inline graphic represent the HL in the decoder and encoder at the time step Inline graphic correspondingly Inline graphic Inline graphic and Inline graphic denote non-linear function activation.

Assuming that Inline graphic is the monitoring deformation data, and Inline graphic is the influencing factor sequence data, with Inline graphic the instance counts and Inline graphic the influencing factor counts, for example, time, temperature, water level, and more. The dual data sequences are first reordered into Inline graphic and Inline graphic respectively, by time steps Inline graphic. The previous output of the HL Inline graphic is then led to the decoder. Every time step of the output decoder is given below:

graphic file with name M77.gif 21

They are associated with sequence to form the last fitting outcome, and the distortion Inline graphic is lastly attained once the denormalization is over.

CAE classifier

The CAE effortlessly incorporates local convolution networks and conventional AEs, presenting a reconstruction feature to the convolutional method33. This feature maps the transformation from input to output called convolutional decoding. Applying the fundamental unsupervised greedy training characteristic of AEs, it is becoming possible to calculate the parameters for either encoder or decoder processes. Now, Inline graphic signifies the convolutional encoder function, whereas Inline graphic represents the decoder counterparts. The input contains feature maps Inline graphic, both from the first layer or the previous opinion. This input includes Inline graphic feature mapping, all spanning a region of l × l pixels. The convolution AE process includes Inline graphic convolutional kernels, making Inline graphic feature mapping within the output layer. When such feature mapping derives from the input layer, Inline graphic describes the input channel counts. However, if they originated from previous layers, Inline graphic represents the complete output feature mapping of that last layer. The dimensions of the convolution kernels stand at Inline graphic × Inline graphic, guaranteeing Inline graphic

The group of parameters Inline graphic describes the learning basics of the convolution AE layer. During this, Inline graphic and Inline graphic be similar to the convolution encoding parameters. Now, every Inline graphic may additionally be characterized as a vector Inline graphic. However, Inline graphic and Inline graphic denotes parameters for the convolution decoding. For this one, Inline graphic and every Inline graphic.

Firstly, the input image experiences an encoder method. In this stage, size patches Inline graphic × Inline graphic pixels signified as Inline graphic while Inline graphic, are removed from the input images. Then, for all patches, the weighting Inline graphic of the Inline graphicth convolutional kernel was applied to convolutional processes. These outcomes within the calculation of the values of the neuron Inline graphic for Inline graphic denotes the output layer:

graphic file with name M107.gif 22

While Inline graphic signifies a non-linear activation function, the ReLU activation function has been applied in this study.

graphic file with name M109.gif 23

After this, the Inline graphic output from the convolutional decoder experiences encoder, where Inline graphic is reconstructed with Inline graphic to yield Inline graphic

graphic file with name M114.gif 24

Next, the convolutional encoder and decoder operations, Inline graphic, are produced for all samples. Deriving from the reconstruction process, Inline graphic patches and all dimensions of Inline graphic × Inline graphic are gained. The cost function can be described as the MSE among the novel patches of the input images Inline graphic Inline graphic and the reconstructed patches Inline graphic Inline graphic. The particular procedure of cost function is offered in Eq. (25), whereas the reconstruction error can be specified in Eq. (26).

graphic file with name M123.gif 25
graphic file with name M124.gif 26

Applying SGD, the errors and weights are refined iteratively, resulting in the optimizer of the convolution AE layer. After training, this enhanced parameter yields the feature mapping, which is forwarded to the following layers.

During this study, the CAE method was accurately calculated using numerous layers, all providing particular functions for decoding and encoding the input data. This method begins using the input layer, which obtains the scalogram, which is then passed over consecutive convolution layers. This layer gradually decreases the image dimensionalities, separating the main features. Ensuing the encoder method, this method changes to the decoder stage. This reconstructed output is essential for classifying and identifying different error states in the pumps.

ICOA-based parameter selection

Eventually, the hyperparameter tuning method is implemented by ICOA to enhance the classification outcomes of ensemble models34. This model was chosen because of its ability to optimize hyperparameters in complex ML models. The ICOA technique improves the standard COA approach by incorporating adaptive mechanisms for improved exploration and exploitation during the search process. Unlike grid or random search, ICOA dynamically navigates the search space, mitigating computation time while achieving more accurate parameter tuning. Its robustness in averting local optima ensures optimal configurations for ensemble models, enhancing accuracy and stability. This method is particularly advantageous for high-dimensional parameter spaces, where conventional techniques often face difficulty with efficiency and precision.

COA is a new metaheuristic intellectual optimizer model. Coati’s behaviours stimulate it, specifically their techniques of hunting and attacking iguanas while avoiding predators, to tackle optimization issues.

Initialization

The below-mentioned formulation signifies the original members of the coati population:

graphic file with name M125.gif 27

Here, Inline graphic signifies the Inline graphicth dimensional location of Inline graphicth coati, Inline graphic denotes a randomly generated actual number within an interval of Inline graphic Inline graphic and Inline graphic indicate the upper and lower bounds in the dimension Inline graphic, respectively; Inline graphic specifies an amount of coati population, Inline graphic represents several sizes.

Hunt and attack tactics

Coatis will search iguanas by climbing trees, and the below-given formulation will signify the coati’s location in the tree:

graphic file with name M136.gif 28

The subsequent formulations express the coati’s location on the ground and then the iguana’s arrival:

graphic file with name M137.gif 29

If the coati’s novel site improves an objective value of a function, then it is accepted; or else, it stays unmoved. This process is stated to as the greedy law and is expressed below:

graphic file with name M138.gif 30

While, Inline graphic signifies the new site of Inline graphicth coati in dimension Inline graphic, Inline graphic refers to a stochastic actual numeral in the range of Inline graphic Inline graphic represents the dimensional Inline graphic location of iguana, signifying the location of an optimum member of the population, Inline graphic is a number which is selected at random in the group of {1, 2}. Inline graphic indicates the Inline graphicth dimensional location of arbitrarily generated iguana under the tree, Inline graphic represents the main function of Inline graphic coati value, and Inline graphic specifies the main value of function of iguana on the base.

Escape from predators

Once a coati challenges a predator, the below-mentioned formulations can signify the random position of the coati’s escape:

graphic file with name M152.gif 31

Inline graphic and Inline graphic signify the local and upper bounds of the Inline graphicth dimension, respectively; Inline graphic denotes the present number of iterations, while Inline graphic specifies the maximum iteration count. Inline graphic refers to the novel location of Inline graphicth coati in Inline graphicth dimension throughout the 2nd phase.

The initialized population quality is essential in the meta-heuristic technique, considerably manipulating both the convergence speed and the accuracy of the last solution. The traditional COA utilizes a randomly generated model for initialization, which leads to a non-uniform spread of solution individuals. Here, the population initialization procedure is enhanced by applying the refractive opposite learning tactic to supplement the model’s performance by enlarging its range of searches.

The refractive index Inline graphic was determined from the regular relationship.

graphic file with name M162.gif 32

Assume that Inline graphic, comprised in the abovementioned formulation and prolonged to the multi-dimensional space, yields the refractive opposite solution Inline graphic:

graphic file with name M165.gif 33

In the COA expansion stage, the coati adapts its site throughout the search procedure as per the present individual optimal, resulting in an early convergence to a local optimal solution by creating its efficiency in global exploration. This study provides a Levy flight method to improve the location upgrade procedure and tackle this limitation.

This technique yields randomized step distances and widens the exploration area, possibly improving the range of the coati population. The improved model for upgrading locations of coati is given below:

graphic file with name M166.gif 34
graphic file with name M167.gif 35

Inline graphic denotes the usual function of Gamma, Inline graphic refers to an arbitrarily produced variable within the interval of Inline graphic Inline graphic signifies the generated variable at random within the range Inline graphic, Inline graphic and Inline graphic obey the usual distributions Inline graphic and Inline graphic, correspondingly. The mathematical formulation is expressed below:

graphic file with name M177.gif
graphic file with name M178.gif 36

The Levy flight approach has been presented to the COA growth stage to improve its global exploration capability and alleviate early convergence. By making longer-distance jumps in the solution space, Levy flight permits a more varied population distribution by enhancing the model’s capability to run away from local goals and efficiently discover the global optimal.

Fitness selection is one of the great factors inducing the outcome of the ICOA approach. The range of the hyperparameter model comprises the solution-encoded system for estimating the efficiency of the candidate solution. Currently, the ICOA approach studies accuracy as the foremost standard for planning FF.

graphic file with name M179.gif 37
graphic file with name M180.gif 38

Where Inline graphic signifies the positive value of true, and Inline graphic represents the positive value of false.

Result analysis and discussion

The MOEM-SMIADP technique’s simulation validation is verified under the HAR dataset35. The dataset contains 10,100 records under six classes, as shown in Table 1. The total number of features is 561, but only 285 are selected.

Table 1.

Details of the HAR dataset.

Class Records
HAR dataset
 Walking 1700
 Upstairs 1500
 Downstais 1400
 Sitting 1700
 Standing 1900
 Laying down 1900
 Total records 10,100

Figure 3 represents the classifier results of the MOEM-SMIADP methodology on the HAR dataset. Figure 3a and b displays the confusion matrices with correct recognition and classification of all classes under 70%TRPH and 30%TSPH. Figure 3c demonstrates the PR analysis, indicating superior performance over all class labels. At the same time, Fig. 3d demonstrates the ROC values, indicating capable results with better ROC analysis for dissimilar classes.

Fig. 3.

Fig. 3

HAR dataset (a,b) confusion matrix, (c) curve of PR, and (d) curve of ROC.

In Table 2 and Fig. 4, the indoor activity detection of the MOEM-SMIADP approach is established on the HAR dataset. The results reported that the MOEM-SMIADP approach correctly discriminated each sample. On 70%TRPH, the MOEM-SMIADP approach offers average Inline graphic of 98.61%, Inline graphic of 95.80%, Inline graphic of 95.75%, Inline graphic of 95.77%, and Inline graphic of 95.77%. Similarly, on 30%TRPH, the MOEM-SMIADP model presents an average Inline graphic of 98.51%, Inline graphic of 95.58%, Inline graphic of 95.57%, Inline graphic of 95.57%, and Inline graphic of 95.77%.

Table 2.

Indoor activity detection of MOEM-SMIADP model on HAR dataset.

Classes Accu y Prec n Reca l Inline graphic Inline graphic
TRPH (70%)
 Walking 98.85 95.62 97.76 96.68 96.68
 Upstairs 98.73 95.74 95.74 95.74 95.74
 Downstais 98.50 95.20 93.60 94.39 94.39
 Sitting 98.76 96.50 96.01 96.26 96.26
 Standing 98.61 96.20 96.62 96.41 96.41
 Laying down 98.20 95.55 94.75 95.15 95.15
 Average 98.61 95.80 95.75 95.77 95.77
TSPH (30%)
 Walking 98.71 95.77 96.36 96.06 96.07
 Upstairs 98.75 95.31 96.17 95.74 95.74
 Downstais 98.51 95.68 94.18 94.93 94.93
 Sitting 98.94 97.48 96.36 96.92 96.92
 Standing 98.51 94.74 97.03 95.87 95.87
 Laying down 97.66 94.46 93.33 93.90 93.90
 Average 98.51 95.58 95.57 95.57 95.57

Fig. 4.

Fig. 4

Average of MOEM-SMIADP model on HAR dataset.

Figure 5 illustrates the training (TRA) Inline graphic and validation (VAL) Inline graphic analysis of the MOEM-SMIADP technique on HAR dataset. The Inline graphic analysis is calculated over the range of 0–50 epochs. The figure highlights that the TRA and VAL Inline graphic analysis shows an increasing tendency, which informed the capacity of the MOEM-SMIADP methodology with maximal outcomes across various iterations. Furthermore, the TRA and VAL Inline graphic leftovers closer across the epochs, which specifies inferior overfitting and demonstrates a higher result of the MOEM-SMIADP method, assuring continuous prediction on hidden samples.

Fig. 5.

Fig. 5

Accuy curve of MOEM-SMIADP model on HAR dataset.

Figure 6 shows the TRA loss (TRALOS) and VAL loss (VALLOS) curves of the MOEM-SMIADP technique on the HAR dataset. The loss values are computed within the range of 0–50 epochs. It is denoted that the TRALOS and VALLOS values exemplify a diminishing tendency, notifying the capacity of the MOEM-SMIADP method in balancing a tradeoff between data fitting and simplification.

Fig. 6.

Fig. 6

Loss curve of MOEM-SMIADP model on HAR dataset.

Table 3 and Fig. 7 compare the outcomes of the MOEM-SMIADP method on the HAR dataset with those of the existing techniques. The outcomes highlight that the CNN, CNN-LSTM, Lightweight CNN, EDA-LSTM, WISNet, MLP, CNN-2D, and Optimized ResNet-34 methodologies have reported lowest performance. Meanwhile, the CNN-BiLSTM method has attained closer outcomes. Simultaneously, the MOEM-SMIADP approach reported maximal performance with lesser Inline graphic, Inline graphic Inline graphic and Inline graphic of 95.80%, 95.75%, 98.61%, and 95.77%, correspondingly.

Table 3.

Comparative analysis of the MOEM-SMIADP model on the HAR dataset.

Techniques Inline graphic Inline graphic Inline graphic Inline graphic
HAR dataset
 CNN classifier 95.31 92.29 90.80 94.76
 CNN-LSTM 95.80 91.59 90.47 95.24
 Lightweight CNN 96.27 94.82 91.40 91.82
 CNN-BiLSTM 97.05 93.14 93.81 90.92
 EDA-LSTM 96.89 91.94 93.74 92.72
 WISNet model 95.66 91.13 95.14 95.40
 MLP method 87.41 92.64 92.64 90.50
 CNN-2D model 88.09 90.52 91.19 93.87
 Optimized ResNet-34 90.80 91.40 90.53 93.95
 MOEM-SMIADP 98.61 95.80 95.75 95.77

Fig. 7.

Fig. 7

Comparative analysis of the MOEM-SMIADP model on HAR dataset.

In Table 4 and Fig. 8, the comparative results of the MOEM-SMIADP method on the HAR dataset are identified in terms of processing time (PT). Based on PT, the MOEM-SMIADP method offers minimal CT of 1.05 s whereas the CCN, CNN-LSTM, Lightweight CNN, CNN-BiLSTM, EDA-LSTM, WISNet, MLP, CNN-2D, and Optimized ResNet-34 approaches achieve superior PT values of 2.53 s, 3.16 s, 5.06 s, 3.52 s, 6.33 s, 3.77 s, 4.52 s, 5.90 s, and 3.98 s, respectively.

Table 4.

PT outcome of MOEM-SMIADP technique on HAR dataset.

Techniques Processing time (sec)
HAR dataset
 CNN classifier 2.53
 CNN-LSTM 3.16
 Lightweight CNN 5.06
 CNN-BiLSTM 3.52
 EDA-LSTM 6.33
 WISNet model 3.77
 MLP method 4.52
 CNN-2D model 5.90
 Optimized ResNet-34 3.98
 MOEM-SMIADP 1.05

Fig. 8.

Fig. 8

PT outcome of MOEM-SMIADP technique on HAR dataset.

Likewise, the performance evaluation of the MOEM-SMIADP technique is verified below the WISDM dataset36. The dataset consists of 30,000 instances below six classes, as shown in Table 5. There are 5 number of features, but only 3 features are selected.

Table 5.

Details of WISDM dataset.

Class No. of instances
WISDM dataset
 Walking 5000
 Jogging 5000
 Upstairs 5000
 Downstais 5000
 Sitting 5000
 Standing 5000
 Total instances 30,000

Figure 9 shows the classifier performances of the MOEM-SMIADP method on the WISDM dataset. Figure 9a and b illustrates the confusion matrices with specific classification and identification of all classes below 70%TRPH and 30%TSPH. Figure 9c exemplifies the PR study, which noted the enhanced performance of all class labels. Eventually, Fig. 9d demonstrates the ROC study, which signifies proficient performances with great ROC values for all dissimilar classes.

Fig. 9.

Fig. 9

WISDM Dataset (a,b) Confusion matrix, (c) curve of PR, and (d) curve of ROC.

In Table 6 and Fig. 10, the indoor activity detection of the MOEM-SMIADP approach is depicted on the WISDM dataset. The performances indicated that the MOEM-SMIADP approach accurately differentiated all the samples. Using 70%TRPH, the MOEM-SMIADP model provides average Inline graphic of 99.06%, Inline graphic of 97.20%, Inline graphic of 97.19%, Inline graphic of 97.19%, and Inline graphic of 97.19%. Additionally, using 30%TRPH, the MOEM-SMIADP method delivers an average Inline graphic of 99.07%, Inline graphic of 97.22%, Inline graphic of 97.22%, Inline graphic of 97.22%, and Inline graphic of 97.22%.

Table 6.

Indoor activity detection of MOEM-SMIADP model on the WISDM dataset.

Classes Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
TRPH (70%)
 Walking 99.19 97.50 97.59 97.54 97.54
 Jogging 98.73 95.39 97.14 96.26 96.26
 Upstairs 99.12 98.12 96.62 97.37 97.37
 Downstais 99.16 97.28 97.72 97.50 97.50
 Sitting 99.05 97.29 96.98 97.13 97.13
 Standing 99.12 97.60 97.06 97.33 97.33
 Average 99.06 97.20 97.19 97.19 97.19
TSPH (30%)
 Walking 99.33 97.90 98.16 98.03 98.03
 Jogging 98.81 95.82 96.93 96.37 96.38
 Upstairs 99.24 98.22 97.16 97.69 97.69
 Downstais 99.04 96.73 97.51 97.12 97.12
 Sitting 99.04 97.24 97.11 97.17 97.17
 Standing 98.97 97.42 96.46 96.94 96.94
 Average 99.07 97.22 97.22 97.22 97.22

Fig. 10.

Fig. 10

Average of MOEM-SMIADP model on WISDM dataset.

Figure 11 depicts the TRA Inline graphic and VAL Inline graphic performances of the MOEM-SMIADP technique on the WISDM dataset. The Inline graphic values are calculated through an interval of 0–50 epochs. The figure implied that the values of TRA and VAL Inline graphic presents a increasing trend, indicating the capacity of the MOEM-SMIADP technique with higher performance across numerous repetitions. In addition, the TRA and VAL Inline graphic values remain close through the epochs, notifying lesser overfitting and displaying the MOEM-SMIADP model’s superior performance, which assurances reliable prediction on unnoticed samples.

Fig. 11.

Fig. 11

Accuy curve of MOEM-SMIADP model on WISDM dataset.

Figure 12 shows the TRALOS and VALLOS graph of the MOEM-SMIADP approach on the WISDM dataset. The loss values are computed across a range of 0–50 epochs. The values of TRALOS and VALLOS show a diminishing trend, which indicates the proficiency of the MOEM-SMIADP model in harmonizing a tradeoff between generalization and data fitting.

Fig. 12.

Fig. 12

Loss curve of MOEM-SMIADP model on WISDM dataset.

Table 7 and Fig. 13 study the comparison results of the MOEM-SMIADP method on the WISDM dataset with the existing techniques3740. The performances indicated that the CNN-LSTM, CNN, CNN-BiLSTM, ALSTM-1D CNN, Mechanism-DL, WISNet, MLP, CNN-2D, and Optimized ResNet-34 models have testified to poorer performance. While, the MOEM-SMIADP model stated maximal performance with higher Inline graphic, Inline graphicInline graphic and Inline graphic of 97.22%, 97.22%, 99.07%, and 97.22%, respectively.

Table 7.

Comparative analysis of the MOEM-SMIADP model on the WISDM dataset3740.

Techniques Inline graphic Inline graphic Inline graphic Inline graphic
WISDM dataset
 CNN-LSTM 95.75 91.18 91.00 93.98
 CNN classifier 97.51 97.06 94.30 92.44
 CNN-BiLSTM 98.53 90.90 94.03 92.01
 ALSTM-1D CNN 98.11 96.29 95.36 96.16
 Mechanism-DL 93.89 96.57 96.87 94.26
 WISNet model 96.41 90.06 90.98 90.87
 MLP method 87.30 92.91 95.22 94.98
 CNN-2D model 94.23 91.24 91.23 93.42
 Optimized ResNet-34 96.33 94.72 93.32 92.33
 MOEM-SMIADP 99.07 97.22 97.22 97.22

Fig. 13.

Fig. 13

Comparative analysis of MOEM-SMIADP model on WISDM dataset.

In Table 8 and Fig. 14, the comparative analysis of the MOEM-SMIADP approach on the WISDM dataset is identified in terms of execution time (ET). According to ET, the MOEM-SMIADP approach presents minimal ET of 4.57 s while the CNN-LSTM, CNN, CNN-BiLSTM, ALSTM-1D CNN, Mechanism-DL, WISNet, MLP, CNN-2D, and Optimized ResNet-34 methodologies obtain better ET values of 7.33 s, 7.45 s, 17.58 s, 11.36 s, 16.21 s, 9.50 s, 9.03 s, 17.15 s, and 9.35 s, respectively.

Table 8.

ET outcome of MOEM-SMIADP technique on WISDM dataset.

Techniques Execution time (sec)
WISDM dataset
 CNN-LSTM 7.33
 CNN classifier 7.45
 CNN-BiLSTM 17.58
 ALSTM-1D CNN 11.36
 Mechanism-DL 16.21
 WISNet model 9.50
 MLP method 9.03
 CNN-2D model 17.15
 Optimized ResNet-34 9.35
 MOEM-SMIADP 4.57

Fig. 14.

Fig. 14

ET outcome of MOEM-SMIADP technique on WISDM dataset.

Conclusion

In this study, a MOEM-SMIADP model is developed. The proposed MOEM-SMIADP model concentrates on detecting and classifying indoor activities using IoT applications for physically challenged people. It encompasses four steps: data normalization, MPA-based feature selection, an ensemble of classification models, and parameter selection using ICOA. At first, the data preprocessing executes min–max normalization to convert input data into useful format. Furthermore, the MPA has been applied to the process of feature selection. For the detection of indoor activities, the proposed MOEM-SMIADP model applies an ensemble of three classifiers: the GCN model, the LSTM-seq2seq method, and the CAE. Eventually, the hyperparameter tuning method is implemented by ICOA to enhance the classification outcomes of ensemble models. A wide range of experiments was accompanied to endorse the performance of the MOEM-SMIADP technique. The performance validation of the MOEM-SMIADP technique portrayed a superior accracy value of 99.07% over existing methods. The limitation of the MOEM-SMIADP technique is its dependence on specific data characteristics, which may not be generalized well to all indoor environments or activity types. The model’s performance might degrade when faced with noisy or incomplete data, as it needs high-quality, well-labelled datasets for optimal results. Furthermore, the computational complexity of the ensemble approach may limit its scalability in real-time applications. Future work may focus on incorporating more robust data preprocessing techniques to handle noise, exploring transfer learning to adapt to diverse environments, and improving the efficiency of the model for deployment in resource-constrained devices. Further investigation into hybrid models incorporating multiple data modalities could also improve activity detection accuracy.

Acknowledgements

The authors extend their appreciation to the King Salman center For Disability Research for funding this work through Research Group no KSRG-2024- 426.

Author contributions

Conceptualization: Munya A. Arasi, Hussah Nasser AlEisa, Data curation and Formal analysis: Amani A Alneil and Radwa MarzoukInvestigation and Methodology: Hussah Nasser AlEisa, Amani A Alneil and Radwa Marzouk, Project administration and Resources: Munya A. Arasi, Writing—original draft: Munya A. Arasi, Hussah Nasser AlEisa, Amani A Alneil and Radwa MarzoukValidation and Visualization: Munya A. Arasi, Hussah Nasser AlEisa, Amani A Alneil and Radwa Marzouk, Writing—review and editing, Munya A. Arasi, Hussah Nasser AlEisa, Amani A Alneil and Radwa MarzoukAll authors have read and agreed to the published version of the manuscript.

Data availability

The data that support the findings of this study are openly available at https://archive.ics.uci.edu/dataset/240/human+activity+recognition+using+smartphones and https://www.cis.fordham.edu/wisdm/dataset.php, reference number35,36.

Declarations

Competing interests

The authors declare no competing interests.

Footnotes

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Abokhoza, R. & Jahmani, A. Towards retention in airline industry using neutrosophic DEMATEL method: Does social media marketing activities affect passengers’ retention. Int. J. Neutrosophic Sci. IJNS21(2), 161–176 (2023). [Google Scholar]
  • 2.Dhiman, C. & Vishwakarma, D. K. A review of state-of-the-art techniques for abnormal human activity recognition. Eng. Appl. Artif. Intell.77, 21–45 (2019). [Google Scholar]
  • 3.Gupta, N. et al. Human activity recognition in artificial intelligence framework: A narrative review. Artif. Intell. Rev.55(6), 4755–4808 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Alotaibi, F. et al. Internet of Things-driven human activity recognition of elderly and disabled people using arithmetic optimization algorithm with LSTM autoencoder. J. Disabil. Res.2(3), 136–146 (2023). [Google Scholar]
  • 5.Perez, A. J., Siddiqui, F., Zeadally, S. & Lane, D. A review of IoT systems to enable independence for the elderly and disabled individuals. Internet Things21, 100653 (2023). [Google Scholar]
  • 6.Rakshanasri, S. L., Naren, J., Vithya, G., Akhil, S. & Kumar, D. A framework on health smart home using IoT and machine learning for disabled people. Int. J. Psychosoc. Rehabil.24(2), 01–09 (2020). [Google Scholar]
  • 7.Brik, B., Esseghir, M., Merghem-Boulahia, L. & Snoussi, H. An IoT-based deep learning approach to analyze indoor thermal comfort of disabled people. Build. Environ.203, 108056 (2021). [Google Scholar]
  • 8.Bibbò, L., Carotenuto, R. & Della Corte, F. An overview of indoor localization system for human activity recognition (HAR) in healthcare. Sensors22(21), 8119 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Lentzas, A. & Vrakas, D. Non-intrusive human activity recognition and abnormal behavior detection on elderly people: A review. Artif. Intell. Rev.53(3), 1975–2021 (2020). [Google Scholar]
  • 10.Arias, E. J., Paz, L. M. A. & Chalacan, L. M. Multi-sensor data fusion for accurate human activity recognition with deep learning. Fusion Pract. Appl.13(2), 62–72 (2023). [Google Scholar]
  • 11.Chen, J., Xu, X., Wang, T., Jeon, G. & Camacho, D. An AIoT framework with multimodal frequency fusion for WiFi-based coarse and fine activity recognition. IEEE Internet Things J.11, 39020–39029 (2024). [Google Scholar]
  • 12.Berkani, M. R. A., Chouchane, A., Himeur, Y., Ouamane, A. & Amira, A. An intelligent edge-deployable indoor air quality monitoring and activity recognition approach. In 2023 6th International Conference on Signal Processing and Information Security (ICSPIS), 184–189 (IEEE, 2023).
  • 13.Sun, H. & Chen, Y. A rapid response system for elderly safety monitoring using progressive hierarchical action recognition. IEEE Trans. Neural Syst. Rehabil. Eng.32, 2134–2142 (2024). [DOI] [PubMed] [Google Scholar]
  • 14.Lee, C., Kang, H. M., Jeon, Y. & Kang, S. J. Ambient sound analysis for non-invasive indoor activity detection in edge computing environments. In 2023 IEEE Symposium on Computers and Communications (ISCC), 1–6 (IEEE, 2023).
  • 15.Mohanaprakash, T. A., Kumar, D., Naveen, P. & Karuppiah, S. Cloud-Enabled Blockchain and IoT-Based Assisted Living System in 6G Networks: Enhancing Quality of Life and Privacy (2024).
  • 16.Kan, R. et al. Indoor human action recognition based on dual kinect V2 and improved ensemble learning method. Sensors23(21), 8921 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Srinivasan, S., Sridevi, V., Saravanan, K., Murugan, S., Srinivasan, C. & Muthulekshmi, M. Adaptive thermal clothing with IoT and random forest regression for dynamic outdoor comfort. In 2024 International Conference on Advances in Modern Age Technologies for Health and Engineering Science (AMATHE), 1–5 (IEEE, 2024).
  • 18.Manimaran, M., Kumar, A. S., Natteshan, N. V. S., Baranitharan, K., Mahaveerakannan, R. & Sudhakar, K. Detecting the human activities of aging people using restricted Boltzmann machines with deep learning technique in IoT. In 2023 Third International Conference on Artificial Intelligence and Smart Energy (ICAIS), 105–110 (IEEE, 2023).
  • 19.Xiao, L., Luo, K., Liu, J. & Foroughi, A. A hybrid deep approach to recognizing student activity and monitoring health physique based on accelerometer data from smartphones. Sci. Rep.14(1), 14006 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Shereef, S., Varghese, N. & Kamalraj, R. Unlocking the power of data: Leveraging IoT and cloud for better sleep health. In Revolutionizing Healthcare Systems Through Cloud Computing and IoT, 179–204 (IGI Global, 2025).
  • 21.Rezaee, K. An advanced deep learning structure for accurate student activity recognition and health monitoring using smartphone accelerometer data. Health Manag. Inf. Sci.11, 85–97 (2024). [Google Scholar]
  • 22.Anitha, A., Nandhini, N., Balakrishnan, K. & Perumal, T. Improving elder care: Vision-based wearable technology for fall recognition and prevention. In Smart Healthcare Systems (eds Bhambri, P. et al.) 304–317 (CRC Press, 2025). [Google Scholar]
  • 23.Maddeh, M. et al. Ensemble learning-based smartbed system for enhanced patient care. J. Disabil. Res.2(1), 26–34 (2023). [Google Scholar]
  • 24.Akhmetshin, E., Nemtsev, A., Shichiyakh, R., Shakhov, D. & Dedkova, I. Evolutionary algorithm with deep learning based fall detection on Internet of Things environment. Fusion Pract. Appl.14(2), 132–145 (2024). [Google Scholar]
  • 25.Namoun, A. et al. Service selection using an ensemble meta-learning classifier for students with disabilities. Multimodal Technol. Interact.7(5), 42 (2023). [Google Scholar]
  • 26.Jawad, M. et al. Energy optimization and plant comfort management in smart greenhouses using the artificial bee colony algorithm. Sci. Rep.15(1), 1752 (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Kao, W. C., Fan, Y. L., Hsu, F. R., Shen, C. Y. & Liao, L. D. Next-generation swimming pool drowning prevention strategy integrating AI and IoT technologies. Heliyon10(18), 1–15 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Yazici, A. et al. A smart e-health framework for monitoring the health of the elderly and disabled. Internet Things24, 100971 (2023). [Google Scholar]
  • 29.Shantal, M., Othman, Z. & Bakar, A. A. A novel approach for data feature weighting using correlation coefficients and min–max normalization. Symmetry15(12), 2185 (2023). [Google Scholar]
  • 30.Hattabi, I. et al. Enhanced power system stabilizer tuning using marine predator algorithm with comparative analysis and real time validation. Sci. Rep.14(1), 28971 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Gao, L. et al. A topology-guided high-quality solution learning framework for security-constraint unit commitment based on graph convolutional network. Int. J. Electr. Power Energy Syst.164, 110322 (2025). [Google Scholar]
  • 32.Wang, L., Wang, J., Tong, D. & Wang, X. A novel long short-term memory seq2seq model with chaos-based optimization and attention mechanism for enhanced dam deformation prediction. Buildings14(11), 3675 (2024). [Google Scholar]
  • 33.Zaman, W., Ahmad, Z. & Kim, J. M. Fault diagnosis in centrifugal pumps: A dual-scalogram approach with convolution autoencoder and artificial neural network. Sensors24(3), 851 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Gong, X. et al. Safety status prediction model of transmission tower based on improved coati optimization-based support vector machine. Buildings14(12), 3815 (2024). [Google Scholar]
  • 35.https://archive.ics.uci.edu/dataset/240/human+activity+recognition+using+smartphones.
  • 36.https://www.cis.fordham.edu/wisdm/dataset.php.
  • 37.Nafea, O., Abdul, W., Muhammad, G. & Alsulaiman, M. Sensor-based human activity recognition with spatio-temporal deep learning. Sensors21(6), 2141 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Sharen, H. et al. WISNet: A deep neural network based human activity recognition system. Expert Syst. Appl.258, 124999 (2024). [Google Scholar]
  • 39.He, Z., Sun, Y. & Zhang, Z. Human activity recognition based on deep learning regardless of sensor orientation. Appl. Sci.14(9), 3637 (2024). [Google Scholar]
  • 40.Khan, I., Guerrieri, A., Serra, E. & Spezzano, G. A hybrid deep learning model for UWB radar-based human activity recognition. Internet Things2024, 101458. 10.1016/j.iot.2024.101458 (2024). [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The data that support the findings of this study are openly available at https://archive.ics.uci.edu/dataset/240/human+activity+recognition+using+smartphones and https://www.cis.fordham.edu/wisdm/dataset.php, reference number35,36.


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES