Summary
Pedestrian navigation activity recognition (PNAR) serves a pivotal role in in the pedestrian positioning and navigation field, providing strong technical support for various aspects such as pedestrian dead reckoning, and multi-source information fusion positioning. This paper proposes a PNAR method that combines a two-stream convolutional transformer architecture with self-supervised contrastive pretraining to address challenges in learning robust, transferable, and generalizable representations from sensor data. The spatial stream captures multi-modal sensor dependencies, while the temporal stream leverages attention mechanism to excavate temporal relationships. The two-stream design effectively processes multi-modal sensor data and models complex activities. Contrastive pretraining leverages unlabeled data to learn invariant and transferable representations, significantly enhancing generalization across datasets. The proposed method was evaluated on four public datasets, achieving exceptional performance—99.08% accuracy and 99.22% F1-score, outperforming existing PNAR methods, including CNNLSTM + Attention and Transformer-based PNAR models. Furthermore, we conducted cross-dataset experiments on data with different sensor configurations and activity labels to validate the model’s superior generalization ability.
Subject areas: computer science, engineering
Graphical abstract

Highlights
-
•
TSCT model temporal and spatial dependencies of inertial sensor data
-
•
We design a Self-supervised contrastive pretraining framework
-
•
Conduct extensive cross-dataset experiments to validate generalization capabilities
Computer science; engineering
Introduction
Human activity recognition (HAR) based on smartphone sensors has been an attractive research topic due to its extensive applications such as smart security,1 healthcare monitoring,2 smart homes,3 and human-computer interaction.4 Pedestrian navigation activity recognition (PNAR) is a subset of HAR.5 In the field of pedestrian positioning and navigation, Accurate recognition of pedestrian navigation activity (such as static, normal walking, running, jumping, upstairs, downstairs, and cycling) is the core technical foundation for achieving high-precision, scene-adaptive navigation and positioning4: (1) motion characteristics compensate for sensor errors. Pedestrian movement patterns such as walking, running, and going up/down stairs exhibit regularity (such as periodic step frequency and step length), which can be used to calibrate the cumulative errors of inertial measurement unit (IMU) through pattern recognition.6 (2) Motion state constrains intelligent fusion of multi-source information.7,8 In multi-source fusion positioning such as IMU/GNSS/Bluetooth/Wi-Fi/vision, motion patterns can serve as state observation measurements while assisting in selecting the optimal positioning source. (3) Enhanced robustness in complex environments. When GNSS signals are blocked in urban canyons, pedestrian dead reckoning mode combined with building layout constraints (such as sidewalk direction) can limit trajectory drift. In indoor scenarios, recognizing user activity patterns of taking elevators/escalators can prevent floor misjudgment (traditional altimeters are susceptible to air pressure fluctuations).9 (4) User intent understanding. Motion patterns can infer user behavioral intentions (such as “stopping to look around” possibly corresponding to interest points), providing contextual information for location-based services (LBSs). In mall navigation, recognizing “wandering” patterns can trigger promotional information. (5) Energy consumption and computational efficiency optimization. Outdoor activities (such as cycling) switch to GPS-dominated mode, reducing unnecessary computation; indoor activities switch to multi-source information fusion mode, improving positioning accuracy; when a stationary state is detected, high-energy modules such as GNSS are automatically turned off while reducing sampling frequency of simple sensors. Therefore, accurate and robust pedestrian navigation activities recognition (PNAR) help design and optimize positioning models to improve positioning accuracy while reducing energy consumption.
According to the different data acquisition devices, current PNAR methods fall into following three major groups: video image-based PNAR methods, radio frequency signal-based PNAR methods, and inertial sensor-based PNAR methods.10,11 The methods based on video images recognizes people’s activity through video and images collected by cameras. This way has the following own defects: one is that the light intensity has a greater impact on the camera. The same activity is different for videos or images under different lighting. At night, the camera cannot even capture any information; the second is that its recognition range is limited. Once the range of the camera is exceeded, the camera cannot collect video or images; the third is that the distance between the recognized person and the camera is different, which increases the difficulty of modeling.12 Radio frequency signal-based activity recognition is to obtain various data such as the azimuth, distance, speed, shape, and size of the target through the electromagnetic waves returned by the radar, thereby completing activity recognition.13 The disadvantages of this method: one is that the price of the radar is relatively expensive; the other is that it is easily affected by other targets and the environment. Inertial sensors can capture people’s motion state very well, without being limited by light, range, and environment. In addition to the rich built-in sensors and strong storage and computing capabilities, the smartphones carried by users can also avoid the disadvantage of being uncomfortable to wear and become the best carrier for ubiquitous activity recognition.14,15
This study focuses on PNAR using inertial sensors embedded in smartphones. Extensive research has been conducted on sensor-based PNAR using classical machine learning models, including decision trees (DTs), K-nearest neighbors (KNNs), hidden Markov models (HMMs), and support vector machines (SVMs). Feature extraction and selection constitute the core components of machine learning-based PNAR methodologies, as they directly influence the discriminative power and generalization capability of the subsequent classification tasks. Deep learning-driven end-to-end PNAR methods automatically extract effective features from raw data, significantly improving recognition accuracy. So far, research on activity recognition is no longer limited to a single network model architecture. It has developed from the simplest machine learning method16,17,18 to simple deep learning network,19,20,21,22 and then used hybrid deep learning networks.23,24,25 Recently, attention-based models have attracted considerable interest in PNAR tasks.26,27,28,29,30,31 Contrastive learning is a self-supervised learning method that helps models to understand sample similarities by comparing pairs of distinct samples. Contrastive learning has shown its powerful representation learning capabilities in many fields. Inspired by the strong advances achieved by contrastive self-supervised learning methods in computer vision (MoCo,32 SimCLR,33 SimSiam,34 and BYOL35), researchers have increasingly adapted these techniques to sensor-based PNAR.36,37,38
Despite the extensive research carried out on activity recognition using smartphone-built inertial sensors, several challenges and difficulties remain unresolved: (1) different human have obvious differences in height, weight, walking habits, etc. The built-in sensors in terminals have many types, varying accuracy, and lack of necessary calibration. Existing activity recognition models struggle to adapt to heterogeneous users and devices39; (2) recognition accuracy is closely related to the device’s posture. There are differences in data distribution between sensor data collected under different location conditions, resulting in the performance of the trained activity recognition model being unable to be guaranteed when facing sensor data collected at other locations.40,41 However, collecting a large amount of labeled data for each location and retraining an activity recognition model will consume a lot of resources and is also unrealistic.42
This paper presents a novel PNAR method by introducing a two-stream convolutional transformer (TSCT) architecture combined with self-supervised contrastive pretraining, addressing key challenges in learning robust, transferable, and generalizable representations from sensor data. The dual-stream design captures spatial dependencies through multi-modal sensor integration and temporal relationships using self-attention mechanisms, enabling effective modeling of complex activity. The contrastive pretraining method leverages unlabeled data, significantly reducing reliance on labeled datasets and improving cross-dataset generalization. The primary contributions of our research are listed below:
-
(1)
We propose a novel architecture to simultaneously model temporal and spatial dependencies in inertial sensor data. The spatial stream captures multi-modal sensor dependencies, while the temporal stream employs an attention mechanism to effectively excavate temporal relationships. This dual-stream design enables the model to comprehensively process complex activities, outperforming single-stream models in capturing diverse dependencies.
-
(2)
We propose a self-supervised contrastive pretraining method based on the Bootstrap Your Own Latent (BYOL) framework to achieve robust PNAR using a small amount of labeled data. We employ a two-stream Transformer as the encoder for BYOL to learn high-level feature representations from unlabeled data during the self-supervised pre-training phase. Subsequently, we utilize a small amount of labeled data for supervised fine-tuning to train a simple linear classifier that classifies the features learned by the encoder, thereby achieving accurate recognition. This method demonstrably enhances model generalization capabilities when applied to cross-dataset scenarios featuring heterogeneous sensor modalities and divergent activity distributions.
-
(3)
We conduct extensive cross-dataset experiments demonstrate the proposed method’s exceptional generalization capacity. Experimental results show consistently superior performance across four public datasets (WISDM, UCI-HAR, PAMAP2, and KU-HAR) with minimal reliance on labeled data. This robust cross-dataset generalization capability, coupled with minimal supervision requirements, underscores the method’s practical viability for real-world positioning and navigation systems.
Results
We carried out comprehensive experiments on public datasets to assess how well the proposed method works. Four widely used public datasets are first described. We detail model parameters. We provide a thorough evaluation of the proposed method’s performance. We trained proposed method on one dataset and tested on another to validate its generalization ability. We evaluate the performance of the proposed method under different scales of labeled data. We reproduced eight baseline models, and conducted comparative experiments.
Datasets for benchmarking
To assess the proposed method’s performance, which uses TSCT and contrastive learning, we conducted experiments on several widely used public datasets. These datasets exhibit significant differences in sensor configuration, activity categories, and physiological characteristics of subjects, which can effectively validate model’s robustness and generalization ability. Table 1 briefly introduces four public datasets, where Acc, Gyr, and Mag represent the accelerometer, gyroscope, and magnetometer, respectively. Figure 1 presents the proportion of different activities in each dataset. The detailed introduction of each dataset is as follows:
Table 1.
Summary of dataset
| DataSets | Frequency | Users | Selected activities | Devices | Sensors | Position | Samples |
|---|---|---|---|---|---|---|---|
| WISDM43 | 20 Hz | 36 | 5 (walking, jogging, upstairs, downstairs, static) | iPhone | Acc | Front pocket of pants | 1098207 |
| UCI-HAR44 | 50 Hz | 30 | 4 (walking, upstairs, downstairs, static) | Galaxy S Ⅱ | Acc, Gyr | Belt | 815614 |
| PAMAP245 | 100 Hz | 9 | 6 (static, walking, running, bicycling, upstairs, downstairs) | 3 Colibri wireless inertial measurement units (IMUs) | Acc, Gyr, Mag | Right arm, chest, and ankle | 2844868 |
| KU-HAR46 | 100 Hz | 90 | 7 (static, forward walking, backward walking, running, jumping, upstairs, downstairs) | Samsung Galaxy J7, Redmi note4, Realme 3 Pro, Realme 5i, Realme C3 | Acc, Gyr | Belt | 6306335 |
Figure 1.
Distribution of proportions of different activities
(A) WISDM.
(B) UCI-HAR.
(C) PAMAP2.
(D) KU-HAR.
WISDM43: this dataset contains data from smartphones carried in users’ pockets during activities like walking, jogging, and stair climbing. The low sampling frequency (20 Hz) and unconstrained device placement make it a challenging dataset, useful for evaluating the model’s adaptability to real-world low-resolution data.
PAMAP245: this dataset contains data from wearable devices positioned on different body parts, capturing a diverse range of 18 activities such as walking, running, cycling, and household chores. The high-frequency data (100 Hz) from accelerometers, gyroscopes, and magnetometers provides rich temporal and spatial features, making it an ideal dataset to evaluate both the spatial and temporal modeling capabilities of the proposed method.
UCI-HAR44: this dataset is widely used for HAR benchmarking, featuring six common activities (e.g., walking, sitting, and standing) recorded using accelerometers and gyroscopes embedded in smartphones. Its relatively small size and simple activity classes make it a suitable baseline for evaluating the core effectiveness of the method.
KU-HAR46: the dataset is a smartphone sensor-based HAR dataset released by Khulna University in 2020. This dataset records 18 activities from 90 participants. Among the 90 participants, 75 were male and 15 were female, with age spanning of 18–34 and an average age of 21.7.
These datasets collectively offer a comprehensive benchmarking environment to test the model’s effectiveness across a spectrum of PNAR challenges, including multimodal integration, temporal dynamics, inter-subject variability, and device heterogeneity.
Model parameters
During the pre-training stage, the TSCT model serves as the encoder, where the deep convolutional blocks employ 64, 128, 256, and 512 filters, respectively, with a uniform kernel size of three. The convolutional feedforward module incorporates 512 filters with kernel sizes of three and one. BYOL is adopted as the contrastive learning framework; both the projector and predictor utilize MLP layers of dimension 512, and the exponential moving average coefficient is set to 0.991. Optimization is performed using Adam with an initial learning rate of 3 × 10−4, which is adaptively scheduled following a simulated cosine annealing strategy. Pre-training relies exclusively on unlabeled data, with each batch comprising 512 samples, and the training process is capped at 50 epochs.
In the subsequent fine-tuning phase, the pre-trained TSCT model is employed solely for feature extraction, upon which a linear classifier is appended. To assess the efficacy of the pre-training, linear evaluation is conducted by freezing all parameters of the TSCT encoder and training only the linear classification layer using a limited set of labeled samples. The initial learning rate is set to 3 × 10−3 and follows the same simulated cosine annealing adjustment policy.
The number of model parameters and the number of FLOPs is 73.22 M and 22.14 M, respectively. This is mainly reflected in model training, but offline training is often cost-insensitive. The online application costs of proposed method are acceptable and can be run on smartphones.
Experiment results on public datasets
To avoid accidental results, we use five repetitive leave-one-subject-out cross-validation (LOSO-CV) tests to evaluate the performance of proposed method. Table 2 provides a comprehensive summary of the proposed method’s performance across four benchmark datasets. The proposed method consistently achieved high classification metrics, with accuracy ranging from 98.62% (KU-HAR) to 99.51% (UCI-HAR). Precision, recall, and F1 score metrics similarly indicate robust recognition capabilities, with all values exceeding 98% across datasets. Notably, the proposed method excelled in datasets with well-separated dynamic activities, such as WISDM (F1 score of 99.21%), while slightly lower performance was observed on datasets with complex activities like KU-HAR (F1 score of 98.26%). The proposed method has good learning ability on four different datasets, and the average recognition accuracy can reach 99.08%. The experimental findings verified the proposed method’s performance, demonstrating its robustness and reliability in PNAR tasks.
Table 2.
Performance metrics on individual datasets
| Dataset | Accuracy (%) | Precision (%) | Recall (%) | F1 score (%) |
|---|---|---|---|---|
| WISDM | 99.32 ± 1.11 | 99.11 ± 1.18 | 99.14 ± 1.21 | 99.12 ± 1.03 |
| UCI-HAR | 99.51 ± 1.23 | 99.25 ± 1.14 | 99.17 ± 1.16 | 99.21 ± 1.08 |
| PAMAP2 | 98.87 ± 1.27 | 98.93 ± 1.23 | 98.62 ± 1.21 | 98.77 ± 1.23 |
| KU-HAR | 98.62 ± 1.38 | 98.39 ± 1.25 | 98.14 ± 1.22 | 98.26 ± 1.29 |
| Average | 99.08 ± 1.25 | 98.92 ± 1.20 | 98.77 ± 1.20 | 99.22 ± 1.16 |
Bold shows the average performance on different datasets.
Figure 2 illustrates the distribution of proportionate predictions across different activity categories for four distinct datasets. These confusion matrices visualize the predictive accuracy of each activity category by displaying the percentage of predictions (rows) that correctly match the actual categories (columns). From the confusion matrix, we can see that the values on the main diagonal are much larger than the rest of the values in the matrix, which shows that the proposed method has outstanding accuracy and effectively distinguish different activity categories.
Figure 2.
Different datasets’ confusion matrix
(A) WISDM.
(B) UCI-HAR.
(C) PAMAP2.
(D) KU-HAR.
Cross-dataset generalization capability validation
We evaluate the cross-dataset generalization capability of the proposed method by training the proposed PNAR model on one dataset and testing it on another dataset. The results shown in Table 3, demonstrate the robustness of the proposed method in generalizing across datasets. When trained on the PAMAP2 and tested on WISDM, the proposed method achieved 97.67% accuracy, 96.88% precision, 97.65% recall, and 97.26% F1 scores. Similarly, when tested on UCI-HAR, the proposed method achieved slightly higher metrics, including 97.89% accuracy and 97.60% F1 score. Notably, when trained on KU-HAR, the proposed method exhibited superior generalization, achieving 98.17% accuracy on WISDM and 98.23% accuracy on UCI-HAR, with F1 scores consistently exceeding 97.30%, highlighting its adaptability to datasets with differing activity distributions and sensor setup. These results underscore the model’s ability to learn invariant and transferable features across datasets with varying complexities and sensor modalities, demonstrating its robustness for real-world activity recognition applications.
Table 3.
Performance metrics on cross-dataset
| Train dataset | Test Dataset | Accuracy (%) | Precision (%) | Recall (%) | F1 score (%) |
|---|---|---|---|---|---|
| PAMAP2 | WISDM | 97.67 | 96.88 | 97.65 | 97.26 |
| UCI-HAR | 97.89 | 97.23 | 97.97 | 97.60 | |
| KU-HAR | WISDM | 98.17 | 97.31 | 97.48 | 97.39 |
| UCI-HAR | 98.23 | 97.95 | 97.69 | 97.82 |
Robustness to annotation scarcity
To evaluate the performance of the proposed method under different scales of labeled data, the proposed model was trained using 2%–20% of labeled data, respectively. Table 4 summarizes the accuracy variations on the WISDM, UCI-HAR, PAMAP2, and KU-HAR datasets. As the amount of labeled data increases from 2% to 20%, the recognition accuracy shows a significant improvement. Due to the smaller sample size of the UCI-HAR dataset, the performance of the proposed method is inferior to the other three datasets when the labeling data ratio is low. The model trained with 12% labeled data achieves recognition accuracies of 99.32%, 99.51%, 98.87%, and 98.62% on the WISDM, UCI-HAR, PAMAP2, and KU-HAR datasets, respectively, indicating that the model has good fast convergence.
Table 4.
Mean accuracy (%) of proposed methods trained by different labeled data on individual datasets
| Dataset | 2% | 4% | 6% | 8% | 10% | 12% | 14% | 16% | 18% | 20% |
|---|---|---|---|---|---|---|---|---|---|---|
| WISDM | 56.34 | 68.65 | 83.24 | 92.32 | 98.91 | 99.32 | 99.33 | 99.36 | 99.35 | 99.37 |
| UCI-HAR | 47.61 | 58.84 | 67.26 | 82.58 | 94.37 | 99.51 | 99.53 | 99.53 | 99.52 | 99.55 |
| PAMAP2 | 57.43 | 71.87 | 89.12 | 95.45 | 98.37 | 98.87 | 98.91 | 98.94 | 98.97 | 99.03 |
| KU-HAR | 62.67 | 86.39 | 93.76 | 97.33 | 98.22 | 98.62 | 98.68 | 98.70 | 98.70 | 98.73 |
| Average | 56.01 | 71.44 | 83.35 | 91.92 | 97.47 | 99.08 | 99.12 | 99.13 | 99.13 | 99.17 |
Bold shows the impact of different training data volumes on performance. Training with 12% labeled data yields good performance, and further augmentation of the training data provides only limited performance improvement.
Comparison with other methods
To assess the advanced performance, we reproduced eight baseline models, which included conventional machine learning methods like KNN and SVM, along with deep learning methods like CNN, LSTM, and transformer. Comparative experiments were performed based on the datasets mentioned earlier. Table 5 summarizes the results, with the top-performing outcomes in each dataset highlighted in bold. Among the baseline models, KNN and SVM exhibited poor classification performance for PNAR task. In contrast, CNN, LSTM, and their combinations demonstrated better classification accuracy. Both transformer-based models and the proposed method achieved excellent recognition results, particularly for dynamic activities like walking, running, and cycling, owing to the temporal stream’s capability to capture dependencies over a long period.
Table 5.
Comparison with baseline methods
| Models | WISDM |
UCI-HAR |
PAMAP2 |
KU-HAR |
Average |
|||||
|---|---|---|---|---|---|---|---|---|---|---|
| Accuracy (%) | F1 (%) | Accuracy (%) | F1 (%) | Accuracy (%) | F1 (%) | Accuracy (%) | F1 (%) | Accuracy (%) | F1 (%) | |
| KNN | 82.88 | 82.65 | 84.24 | 84.12 | 83.21 | 82.87 | 81.24 | 81.76 | 82.89 | 82.85 |
| SVM | 83.23 | 83.34 | 85.26 | 85.38 | 84.37 | 84.42 | 82.22 | 82.51 | 83.77 | 83.91 |
| Stack-HAR16 | 89.33 | 88.87 | 90.12 | 90.35 | 88.97 | 88.77 | 88.34 | 88.31 | 89.23 | 89.45 |
| SDAE+LightGBM47 | 95.41 | 95.33 | 95.38 | 95.57 | 93.45 | 93.41 | 91.23 | 92.21 | 93.87 | 94.13 |
| CNN48 | 95.13 | 95.21 | 95.87 | 95.99 | 94.89 | 94.94 | 92.34 | 92.89 | 94.56 | 94.76 |
| LSTM49 | 95.78 | 96.12 | 96.83 | 96.89 | 95.41 | 95.72 | 94.76 | 94.89 | 95.69 | 95.90 |
| CNNLSTM50 | 96.98 | 97.31 | 97.46 | 96.41 | 96.32 | 96.54 | 95.11 | 95.22 | 96.47 | 96.37 |
| CNNLSTM + Attention27 | 97.41 | 98.28 | 98.21 | 98.47 | 96.66 | 96.87 | 95.88 | 95.92 | 97.04 | 97.39 |
| Transformer31 | 98.17 | 98.21 | 98.54 | 98.52 | 98.01 | 98.17 | 97.87 | 97.92 | 98.14 | 98.20 |
| CSSHAR51 | 98.22 | 98.29 | 98.87 | 98.92 | 98.41 | 98.57 | 98.03 | 98.18 | 98.38 | 98.49 |
| Proposed | 99.32 | 99.12 | 99.51 | 99.21 | 98.87 | 98.77 | 98.62 | 98.26 | 99.08 | 99.22 |
Bold shows the performance of the proposed method on different datasets.
The proposed method, utilizing a dual-stream transformer structure (comprising a time stream and a spatial stream), demonstrated a 1% accuracy improvement compared to the single-stream transformer structure (time stream alone). These findings highlight the proposed method’s strength and efficiency in learning both spatial and temporal stream features, enabling superior generalization across diverse datasets. The consistent improvements over traditional methods, including KNN, SVM, and hybrid deep learning models, validate the proposed method’s advantages.
Discussion
With breakthroughs in technologies such as flexible electronics and biosensors, PNAR is expected to intersect with multiple disciplines such as genomics and metabolomics to promote the further development of precision medicine. This study introduces an innovative PNAR method that merges a TSCT with self-supervised contrastive pretraining to address the challenges of limited robustness and generalization posed by complex and variable activities. The proposed method effectively captures both spatial and temporal dependencies, enabling it to handle diverse and complex activity patterns across multiple datasets. The integration of self-supervised contrastive pretraining further reduced the reliance on large labeled datasets, enabling the model to effectively utilize unlabeled data and improve transferability across domains.
Extensive experiments on multiple heterogeneous datasets, such as WISDM, UCI-HAR, PAMAP2, and KU-HAR, confirmed the method’s superiority in accuracy, precision, recall, and F1 scores. Furthermore, cross-dataset evaluations demonstrated its exceptional generalization capability, demonstrating it suitability for diverse and significantly varying data distributions.
In the future, the proposed method can be enhanced by incorporating domain adaptation techniques to address more significant domain shifts and multi-task learning approaches to improve robustness. Furthermore, investigating the model’s scalability for real-time deployment and its performance in multi-sensor fusion scenarios would expand its practical applicability.
Limitations of the study
Although the proposed TSCT combined with self-supervised contrastive learning exhibits good PNAR performance, there are still several limitations that need to be addressed in the future. This paper performs performance evaluation on multiple public datasets. Although these datasets cover a variety of activity types, their environments and application scenarios may differ from actual complex environments. Although the method proposed in the paper performs well in terms of accuracy, the model introduces relatively high computational complexity and model scale. Although the cost of online inference is acceptable for smartphones, the model may still face deployment challenges on ultra-low-power wearable devices or edge devices with strict memory and energy consumption constraints. The model’s computational efficiency needs to be further optimized to reduce the model’s inference time and storage requirements to ensure its feasibility in practical applications.
Resource availability
Lead contact
Further information and requests for resources, data, and code should be directed to and will be fulfilled by the lead contact, Jianquan Wang (wangjianquan@ustb.edu.cn).
Materials availability
This study did not generate new materials.
Data and code availability
-
•
All data generated or analyzed during this study are included in this published article.
-
•
This study does not report original code.
-
•
Any additional information required to reanalyze the data reported in this study is available from the lead contact upon request.
Acknowledgments
This work was supported in part by Science and Technology Innovation Program of Xiongan New Area (2025XAGG0028), National Natural Science Foundation of China (42401521, U25A20433, 92567203), Joint Research Fund for Beijing Natural Science Foundation and Haidian Original Innovation (L232001), Henan Key Research and Development Program (241111320700), Guangdong Basic and Applied Basic Research Foundation (2024A1515011866, 2024A1515011480, and 2025A1515011300), Central Guidance on Local Science and Technology Development Fund of ShanXi Province (YDZJSX20231D005, and YDZJSX20231B017), S&T Program of Hebei (246Z0303G).
Author contributions
Conceptualization, Q.W., J.W., and Z.J.; methodology, Q.W., J.M., and Y.B.B.; software, M.F. and Y.B.B.; investigation, J.M., Z.J., and H.L.; writing – original draft, Q.W., J.M., and Y.B.B.; writing – review and editing, all authors; visualization, M.F.; supervision, J.W., and H.L.; project administration, J.W. and M.F.; funding acquisition, Z.J. and H.L. All authors have read and agreed to the published version of the manuscript.
Declaration of interests
The authors declare no competing interests.
Declaration of generative AI and AI-assisted technologies in the writing process
The authors acknowledge the use of ChatGPT and Grammarly for language editing. After using this tool, the authors reviewed and edited the content as needed and take full responsibility for the content of the publication.
STAR★Methods
Key resources table
| REAGENT or RESOURCE | SOURCE | IDENTIFIER |
|---|---|---|
| Software and algorithms | ||
| Python 3.12.3 | Python Software Foundation | https://www.python.org |
| BYOL | Contrastive frameworks | https://github.com/google-deepmind/deepmind-research/tree/master/byol |
Method details
Problem modeling
Smartphones equipped with multiple sensors gather data from various sources. Thus, a sample x is represented as x = {x(1),x(2), …,x(n)}, where n denotes the complete count of directions across all sensors. For instance, if an accelerometer and a gyroscope are used, n = 6. Each x(i) corresponds to the data collected by a sensor in a specific direction. Since sensors typically sample at equal intervals, x(i) is further represented as x(i) = {x(i,1),x(i,2), …,x(i,l)}. The data collected in a single direction from one sensor forms a one-dimensional time series of length l. Consequently, the overall sample is a multi-dimensional time series with a length of l and a dimension n, denoted as . PNAR is a typical classification problem. We assume that the activity sample space is , and the activity category space is Y, with |Y| representing the number of activity categories. Thus, an activity sample corresponds to an activity category . The training set is represented as . The aim is to establish a mapping function f:X→Y using , so that f aligns as consistent as possible with the real mapping relationship freal, namely f ˜ freal.
System architecture
We present a PNAR method, depicted in Figure 3, is based on two-stream transformer and contrastive learning, which consists of three key modules: pre-training module, fine-tuning module, and online testing module, arranged sequentially from bottom to top. The pre-training module employs contrastive learning to train a deep feature encoder and mapper, transforming sensor signals into latent space representations. The module minimizes contrastive loss to optimize the latent space representation, clustering similar category features and dispersing different ones. Through this process, the encoder can effectively capture critical features from the input data, which are well-suited for downstream tasks. Finally, the online testing module applies the frozen encoder and the trained classifier from the fine-tuning stage to classify unlabeled sensor window signals, effectively identifying the corresponding activity categories.
Figure 3.
Proposed PNAR method based on two-stream transformer and contrastive learning
Data pre-processing and augmentation
Data collected by sensors often contains data noise from hardware devices and unrelated activities. In addition, subtle vibrations produced involuntarily during activities may also affect the accuracy of the network model. Therefore, we use fourth-order Butterworth bandpass filtering to remove attenuated data Direct Current (DC) drift, data artifacts due to motion, and other low- and high-frequency noise.
Sampling rates vary between different sensors and different datasets. To unify these signals into a multidimensional matrix as input to the model, we convert all signals to a consistent sampling frequency by upsampling or downsampling. This article unifies the sampling rates of different sensors and datasets to 100Hz. In addition, due to differences in units and dimensions between different sensing data, we perform maximum-minimum normalization on the data.
Sliding window is a commonly used data segmentation method. The window’s size and overlap significantly impact mode’s predictive performance. This paper uses multi-scale sliding windows for data segmentation, in which samples are obtained by sliding under the time dimension of small-scale windows, while retaining the characteristics of large-scale time windows that maintain non-repetitive motion periodic activities, thereby obtaining richer feature.
In the temporal stream, since the self-attention mechanism has difficulty exploiting the sequential correlations of time steps, some temporally sensitive actions are difficult to recognize. Positional Encoding (PE) is added to the nonlinear transformation of the temporal stream sequence data to focus on the temporal sequence information of the data. Since the signals from each sensor axis contribute differently to the recognition results, we add sensor attention to the spatial stream sequence data to focus on the contribution of each sensor axis within the input time window.
Two-stream convolutional transformer model
To derive implicit high-level features, we propose a Two-Stream Convolutional Transformer (TSCT) model, as depicted in Figure 4. The spatial stream captures multi-modal sensor dependencies, while the temporal stream leverages attention mechanism to excavate temporal relationships. The TSCT model consists of five layers: input, preprocessing, encoder, aggregation, and prediction layer. When employed as an encoder in a contrastive self-supervised learning module, the TSCT model includes only the first four layers, excluding the prediction layer. The input layer processes sensor window data independently and feeds it into the dual-stream channels. The Transformer encoder layer extracts distinct features from each stream, which are then integrated in the aggregation layer using global attention mechanisms. Finally, the prediction layer produces the final classification results. The following sections will offer a d etailed explanation of each layer.
Figure 4.
Proposed TSCT model
Transformer encoder layer
The Transformer encoder layer includes a multi-head attention module with a residual structure and a convolutional feedforward module. The Transformer encoder layer utilizes attention mechanism to capture global information from sensor data. The multi-head attention module calculates the similarities between time steps and allocates weights according to the sequence’s internal correlations. This layer computes dependencies of all time steps within a time window, and captures the overall sensor information. First, the time series features generated by preprocessing layer are linearly transformed to produce three matrices: Q as the query matrix, K as key matrix, and V as value matrix:
| (Equation 1) |
where and represent weight matrices that can be learned in linear transformation, respectively. where X is the layer input (dimension for spatial stream, for temporal stream). Vector similarity is calculated by taking the dot product of Q and K. Subsequently, scale QK⊤and SoftMax function are applied to generate attention scores for the corresponding time step. Finally, the attention score is multiplied by the value (V) for weighting operation. The overall calculation process is as follows:
| (Equation 2) |
where dk denotes the size of query matrix Q and key matrix K. dν denotes the size of value matrix V. The input of the temporal stream channel is denoted as X, while the input of the spatial stream channel is denoted as Y.
To enhance the model’s capacity to concentrate on diverse areas, we execute h attention modules in parallel. We use different weight matrices when performing linear transformations to produce multiple Q, K, and V. Each attention module computes its output as Zi = Attention(Qi,Ki,Vi). Finally, the multi-head attention module concatenates the outputs from each attention head and applies another linear transformation as follow:
| (Equation 3) |
where is the corresponding projection matrix. To mitigate the vanishing gradient problem caused by module stacking and preserve information integrity, the attention module incorporates residual connections, ensuring that d = dout.
We combine the attention module’s output with the input sequence and perform layer normalization (LN) operation before passing it to the feedforward layer containing residual connections. The result is recorded as Oa:
| (Equation 4) |
In the standard Transformer architecture, the feedforward layer includes two fully connected layers. To enhance local information extraction, we introduce a convolutional feedforward network (CFN) module. CFN consists of two one-dimensional convolutional layers, each equipped with batch normalization (BN), ReLU activation, and dropout to mitigate overfitting. Its output is recorded as Of. The complete CFN computation is described as follows:
| (Equation 5) |
where W1 and W2 are learnable weight that are convolved twice. We add CFN module’s output Ofto input sequence Oa and perform layer normalization to obtain the final features of the Transformer encoder layer, denoted as Xe:
| (Equation 6) |
Due to the use of residual connections in the Transformer encoder layer, the dimensions of the extracted features remain unchanged.
Aggregation layer
Aggregation layer processes features extracted from the spatial stream channel (with dimensions ) and temporal stream channel (with dimensions ). To facilitate effective feature fusion, a global attention mechanism is utilized to enhance relevant information in the extracted features while suppressing irrelevant details. First, linear projection is performed on Xe to calculate position-specific weights. Then These weights are normalized using the SoftMax function. Finally, we calculate the weighted feature matrix C through the Hadamard product as follows:
| (Equation 7) |
The attention output is produced by multiplying the feature weight matrices with the respective hidden vectors.
| (Equation 8) |
To leverage the complementarity of temporal and spatial information, we concatenate spatial and temporal features into a unified feature vector as the output of aggregation layer U:
| (Equation 9) |
where FS represents the spatial stream feature vector, while FT denotes the temporal stream feature vector.
Output layer
This study constructs the TSCT as a complete model that allows recognition results to be obtained directly from sensor data. Because PNAR is a typical multi-class classification task, its prediction layer is a fully connected layer with SoftMax unit. The fully connected layer adjusts the classification result, allowing SoftMax units to predict the activity categories. We use the standard cross entropy loss function to calculate the distance between predicted value and actual value yi:
| (Equation 10) |
where ∣L∣ represents the overall count of categories. f(·) represents the distribution of prediction. XT is the inputs of temporal stream and XS is the inputs of spatial stream.
Self-supervised contrastive pretraining method
Pretraining is crucial for improving model generalization, especially when labeled data is limited or when there are significant differences between datasets. MoCo32 reformulated several prior contrastive learning methods as a dictionary lookup problem. By utilizing a queue and treating multiple elements within it as negative samples, this approach mitigates computational overhead. BYOL35 directly removes negative samples from MoCo, setting up two distinct networks: an online network and a target network. After obtaining features through the projection head, the online network adds an additional prediction layer composed of one or two fully connected layers. This prediction layer is then used to predict the features obtained by the target network, effectively performing a regression task. The model employs MSE as its loss function. Therefore, we design a self-supervised contrastive pretraining method under BYOL framework that leverages unlabeled data to develop a robust feature encoder, which effectively captures the inherent similarities and differences.
Contrastive loss function
The model employs a normalized cross-entropy loss function to optimize encoder. For each batch of samples , the loss is calculated as follows:
| (Equation 11) |
where zi and zj are the latent representations of the augmented input samples, sim(·,·) denotes cosine similarity, and τ is a temperature parameter.
Encoder design
The TSCT model serves as the encoder in the pretraining stage, leveraging its spatial and temporal streams to extract comprehensive features. Encoders effectively mine local and global features from input signals through self-attention mechanisms.
Supervised fine-tuning
After pre-training, the encoder containing the TSCT architecture is fine-tuned under supervised conditions to adapt to specific PNAR tasks. This phase ensures that the pretrained features are aligned with the target dataset’s label distribution and optimized for classification. A simple fully connected layer, often referred to as a linear classifier, is appended to the output of the pretrained encoder. This classifier projects the latent feature representations into the label space corresponding to the activity classes. For datasets with C activity classes, the output layer has C units, followed by a softmax activation function for probability estimation.
| (Equation 12) |
where z represents the latent representation derived from the pretrained encoder. W and b are weight matrix and bias vector, respectively. The fine-tuning phase optimizes the cross-entropy loss function:
| (Equation 13) |
where yi,c represents the actual label encoded in one-hot format and denotes the estimated likelihood for class c.
Quantification and statistical analysis
We completed quantitative analysis based on Python (PyTorch framework), and conducted model training and evaluation on workstations equipped with NVIDIA GPUs. We use Accuracy, Precision, Recall and F1-score as the main evaluation indicators. To avoid accidental results, we use 5 repetitive Leave-One-Subject-Out Cross-Validation (LOSO-CV) tests to evaluate the performance of the proposed method, and the final results are presented as “mean ± standard deviation”. We use confusion matrix to visually analyze the recognition accuracy on different datasets. In the cross-dataset experiment, after the model was trained on one dataset, it was tested on another dataset without additional domain adaptation to verify the generalization ability of the proposed method under different sensor configurations and activity distribution conditions. To evaluate the performance of the proposed method under different scales of labeled data, the proposed model was trained using 2%–20% of labeled data, respectively. Table 4 summarizes the accuracy variations on the WISDM, UCI-HAR, PAMAP2, and KU-HAR datasets.
Published: April 1, 2026
Contributor Information
Jianquan Wang, Email: wangjianquan@ustb.edu.cn.
Hongdun Li, Email: lihd@motcats.ac.cn.
References
- 1.Hussain A., Khan S.U., Khan N., Ullah W., Alkhayyat A., Alharbi M., Baik S.W. Shots segmentation-based optimized dual-stream framework for robust human activity recognition in surveillance video. Alexandria Eng. J. 2024;91:632–647. doi: 10.1016/j.aej.2023.11.017. [DOI] [Google Scholar]
- 2.Ahmed S., Cho S.H. Machine Learning for Healthcare Radars: Recent Progresses in Human Vital Sign Measurement and Activity Recognition. IEEE Commun. Surv. Tutorials. 2024;26:461–495. doi: 10.1109/COMST.2023.3334269. [DOI] [Google Scholar]
- 3.Ding J., Wang Y., Xie Q., Niu J. A Multiple Wi-Fi Sensors Assisted Human Activity Recognition Scheme for Smart Home. IEEE Sens. J. 2025;25:4958–4968. doi: 10.1109/JSEN.2024.3511087. [DOI] [Google Scholar]
- 4.Qi W., Xu X., Qian K., Schuller B.W., Fortino G., Aliverti A. A Review of AIoT-Based Human Activity Recognition: From Application to Technique. IEEE J. Biomed. Health Inform. 2025;29:2425–2438. doi: 10.1109/JBHI.2024.3406737. [DOI] [PubMed] [Google Scholar]
- 5.Wang J., Chen Y., Hao S., Peng X., Hu L. Deep learning for sensor-based activity recognition: A survey. Pattern Recognit. Lett. 2019;119:3–11. doi: 10.1016/j.patrec.2018.02.010. [DOI] [Google Scholar]
- 6.Jin Z., Zhang X., Liu G., Guo M., Su Y., Lu M. Flexible Gaits-Adaptive Pedestrian Dead Reckoning System: Precision Positioning Across Diverse Gaits. IEEE Sens. J. 2025;25:15431–15441. doi: 10.1109/JSEN.2025.3545833. [DOI] [Google Scholar]
- 7.Wang Q., Fu M., Wang J., Luo H., Sun L., Ma Z., Li W., Zhang C., Huang R., Li X., et al. Recent Advances in Pedestrian Inertial Navigation Based on Smartphone: A Review. IEEE Sens. J. 2022;22:22319–22343. doi: 10.1109/JSEN.2022.3213836. [DOI] [Google Scholar]
- 8.Niu Z., Cong L., Qin H., Cao S. Pedestrian Dead Reckoning Based on Complex Motion Mode Recognition Using Hierarchical Classification. IEEE Sens. J. 2024;24:4935–4947. doi: 10.1109/JSEN.2023.3347228. [DOI] [Google Scholar]
- 9.Wang Q., Fu M., Wang J., Luo H., Sun L., Ma Z., Li W., Zhang C., Huang R., Li X., et al. Recent advances in floor positioning based on smartphone. Measurement. 2023;214 doi: 10.1016/j.measurement.2023.112813. [DOI] [Google Scholar]
- 10.Yin Y., Xie L., Jiang Z., Xiao F., Cao J., Lu S. A Systematic Review of Human Activity Recognition Based on Mobile Devices: Overview, Progress and Trends. IEEE Commun. Surv. Tutorials. 2024;26:890–929. doi: 10.1109/COMST.2024.3357591. [DOI] [Google Scholar]
- 11.Kaur H., Rani V., Kumar M. Human activity recognition: A comprehensive review. Expert Syst. 2024;41:e13680. doi: 10.1111/exsy.13680. [DOI] [Google Scholar]
- 12.Chen K., Zhang D., Yao L., Guo B., Yu Z., Liu Y. Deep Learning for Sensor-Based Human Activity Recognition: Overview, Challenges and Opportunities. ACM Comput. Surv. 2022;54:1–40. doi: 10.1145/3447744. [DOI] [Google Scholar]
- 13.Yang C., Wang X., Mao S. TARF: Technology-Agnostic RF Sensing for Human Activity Recognition. IEEE J. Biomed. Health Inform. 2023;27:636–647. doi: 10.1109/JBHI.2022.3175912. [DOI] [PubMed] [Google Scholar]
- 14.Chen J., Wang Z., Xiao K., Ferraro M., Ushakov N., Kumar S., Ge F., Li X., Min R. AI-Enabled Scalable Smartphone Photonic Sensing System for Remote Healthcare Monitoring. IEEE Internet Things J. 2025;12:4510–4524. doi: 10.1109/JIOT.2024.3485614. [DOI] [Google Scholar]
- 15.Sun Z., Ke Q., Rahmani H., Bennamoun M., Wang G., Liu J. Human Action Recognition From Various Data Modalities: A Review. IEEE Trans. Pattern Anal. Mach. Intell. 2023;45:3200–3225. doi: 10.1109/TPAMI.2022.3183112. [DOI] [PubMed] [Google Scholar]
- 16.Dahal A., Moulik S., Mukherjee R. Stack-HAR: Complex Human Activity Recognition With Stacking-Based Ensemble Learning Framework. IEEE Sens. J. 2025;25:16373–16380. doi: 10.1109/JSEN.2025.3549311. [DOI] [Google Scholar]
- 17.Zhang K., Wang Q., Meng X., Wang J. A Human Activity Recognition Scheme Using Mobile Smartphones Based on Varying Orientations and Positions. IEEE Sens. J. 2024;24:17127–17139. doi: 10.1109/JSEN.2024.3380076. [DOI] [Google Scholar]
- 18.Qiu S., Zhao H., Jiang N., Wang Z., Liu L., An Y., Zhao H., Miao X., Liu R., Fortino G. Multi-sensor information fusion based on machine learning for real applications in human activity recognition: State-of-the-art and research challenges. Inf. Fusion. 2022;80:241–265. doi: 10.1016/j.inffus.2021.11.006. [DOI] [Google Scholar]
- 19.Du Q., Wang Z., Kuang Y., Yao Y., Cao Q., Yang Y. P2Net: A Two-Stage Personalized Pedestrian Dead Reckoning Based on Neural Networks. IEEE Sens. J. 2025;25:5757–5768. doi: 10.1109/JSEN.2024.3515173. [DOI] [Google Scholar]
- 20.Cemiloglu A., Akay B. Handling heterogeneity in Human Activity Recognition data by a compact Long Short Term Memory based deep learning approach. Eng. Appl. Artif. Intell. 2025;153 doi: 10.1016/j.engappai.2025.110788. [DOI] [Google Scholar]
- 21.Su J., Liao Z., Sheng Z., Liu A.X., Singh D., Lee H.-N. Human Activity Recognition Using Self-Powered Sensors Based on Multilayer Bidirectional Long Short-Term Memory Networks. IEEE Sens. J. 2023;23:20633–20641. doi: 10.1109/JSEN.2022.3195274. [DOI] [Google Scholar]
- 22.Khodabandelou G., Chibani A., Amirat Y., Tanimoto S. A Recurrent Neural Network Optimization Method for Anticipation of Hierarchical Human Activity. IEEE Trans. Autom. Sci. Eng. 2024;21:4657–4673. doi: 10.1109/TASE.2023.3300821. [DOI] [Google Scholar]
- 23.Choudhury N.A., Soni B. An Adaptive Batch Size-Based-CNN-LSTM Framework for Human Activity Recognition in Uncontrolled Environment. IEEE Trans. Industr. Inform. 2023;19:10379–10387. doi: 10.1109/TII.2022.3229522. [DOI] [Google Scholar]
- 24.Wang C., Luo H., Zhao F., Qin Y. Combining Residual and LSTM Recurrent Networks for Transportation Mode Detection Using Multimodal Sensors Integrated in Smartphones. IEEE Trans. Intell. Transp. Syst. 2021;22:5473–5485. doi: 10.1109/TITS.2020.2987598. [DOI] [Google Scholar]
- 25.Mahmud T., Akash S.S., Fattah S.A., Zhu W.-P., Ahmad M.O. 2020 IEEE 63rd International Midwest Symposium on Circuits and Systems (MWSCAS) IEEE; 2020. Human Activity Recognition From Multi-modal Wearable Sensor Data Using Deep Multi-stage LSTM Architecture Based on Temporal Feature Aggregation; pp. 249–252. [DOI] [Google Scholar]
- 26.Khatun M.A., Yousuf M.A., Ahmed S., Uddin M.Z., Alyami S.A., Al-Ashhab S., Akhdar H.F., Khan A., Azad A., Moni M.A. Deep CNN-LSTM With Self-Attention Model for Human Activity Recognition Using Wearable Sensor. IEEE J. Transl. Eng. Health Med. 2022;10:2700316. doi: 10.1109/JTEHM.2022.3177710. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Murahari V.S., Plötz T. Proceedings of the 2018 ACM International Symposium on Wearable Computers. ACM; 2018. On attention models for human activity recognition; pp. 100–103. [DOI] [Google Scholar]
- 28.Yang X., Xu Z., Liu H., Shull P.B., Redmond S., Liu G., Wang C. A Multiscale Cross-Modal Interactive Fusion Network for Human Activity Recognition Using Wearable Sensors and Smartphones. IEEE Internet Things J. 2024;11:27139–27152. doi: 10.1109/JIOT.2024.3400022. [DOI] [Google Scholar]
- 29.Tokas P., Semwal V.B., Verma S. A Real-Time Deployable Attention-Driven CNN–LSTM Framework for Human Activity Recognition Using Wearable Sensor. IEEE Sens. J. 2025;25:40461–40471. doi: 10.1109/JSEN.2025.3610667. [DOI] [Google Scholar]
- 30.Tan T.H., Chang Y.L., Wu J.R., Chen Y.F., Alkhaleefah M., Alkhaleefah M. Convolutional Neural Network With Multihead Attention for Human Activity Recognition. IEEE Internet Things J. 2024;11:3032–3043. doi: 10.1109/JIOT.2023.3294421. [DOI] [Google Scholar]
- 31.Wang Q., Tao Z., Ning J., Jiang Z., Guo L., Luo H., Wang H., Men A., Cheng X., Zhang Z. Pedestrian Navigation Activity Recognition Based On Segmentation Transformer. IEEE Internet Things J. 2024;11:26020–26032. doi: 10.1109/JIOT.2024.3394050. [DOI] [Google Scholar]
- 32.He K., Fan H., Wu Y., Xie S., Girshick R. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) IEEE; 2020. Momentum Contrast for Unsupervised Visual Representation Learning; pp. 9726–9735. [DOI] [Google Scholar]
- 33.Chen T., Kornblith S., Norouzi M., Hinton G. A Simple Framework for Contrastive Learning of Visual Representations. 37th Int. Conf. Mach. Learn. ICML. 2020;2020:1597–1607. [Google Scholar]
- 34.Chen X., He K. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. 2021. Exploring simple Siamese representation learning; pp. 15750–15758. [DOI] [Google Scholar]
- 35.Grill J.B., Strub F., Altché F., Tallec C., Richemond P.H., Buchatskaya E., Doersch C., Pires B.A., Guo Z.D., Azar M.G., et al. Advances in Neural Information Processing Systems. 2020. Bootstrap your own latent a new approach to self-supervised learning; pp. 21271–21284. [Google Scholar]
- 36.Taghanaki S.R., Rainbow M., Etemad A. Self-Supervised Human Activity Recognition With Localized Time-Frequency Contrastive Representation Learning. IEEE Trans. Hum. Mach. Syst. 2023;53:1027–1037. doi: 10.1109/THMS.2023.3303438. [DOI] [Google Scholar]
- 37.Taghanaki S.R., Rainbow M., Etemad A., Etemad A., Member S. Self-Supervised Human Activity Recognition With Representation Learning. IEEE Trans. Hum. Mach. Syst. 2023;53:1027–1037. doi: 10.1109/THMS.2023.3303438. [DOI] [Google Scholar]
- 38.Sun Y., Xu X., Tian X., Zhou L., Li Y. Efficient human activity recognition: A deep convolutional transformer-based contrastive self-supervised approach using wearable sensors. Eng. Appl. Artif. Intell. 2024;135 doi: 10.1016/j.engappai.2024.108705. [DOI] [Google Scholar]
- 39.Wang Q., Luo H., Wang J., Sun L., Ma Z., Zhang C., Fu M., Zhao F. Recent Advances in Pedestrian Navigation Activity Recognition: A Review. IEEE Sensors J. 2022;22:7499–7518. [Google Scholar]
- 40.Chang Y., Mathur A., Isopoussu A., Song J., Kawsar F. A Systematic Study of Unsupervised Domain Adaptation for Robust Human-Activity Recognition. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020;4:1–30. doi: 10.1145/3380985. [DOI] [Google Scholar]
- 41.Chen R., Member G.S., Luo H. A Lightweight Deep Human Activity Recognition Algorithm Using Multiknowledge Distillation. IEEE Sens. J. 2024;24:31495–31511. doi: 10.1109/JSEN.2024.3443308. [DOI] [Google Scholar]
- 42.Wei B., Yi C., Zhang Q., Zhu H., Zhu J., Jiang F. ActiveSelfHAR: Incorporating Self-Training Into Active Learning to Improve Cross-Subject Human Activity Recognition. IEEE Internet Things J. 2024;11:6833–6847. doi: 10.1109/JIOT.2023.3314150. [DOI] [Google Scholar]
- 43.Kwapisz J.R., Weiss G.M., Moore S.A. Activity recognition using cell phone accelerometers. SIGKDD Explor. Newsl. 2011;12:74–82. doi: 10.1145/1964897.1964918. [DOI] [Google Scholar]
- 44.Anguita D., Ghio A., Oneto L., Parra X., Reyes-Ortiz J.L. ESANN 2013 proceedings, 21st European Symposium on Artificial Neural Networks. Computational Intelligence and Machine Learning; 2013. A public domain dataset for human activity recognition using smartphones. [Google Scholar]
- 45.Reiss A., Stricker D. 2012 16th International Symposium on Wearable Computers. IEEE; 2012. Introducing a New Benchmarked Dataset for Activity Monitoring; pp. 108–109. [DOI] [Google Scholar]
- 46.Sikder N., Nahid A.-A. KU-HAR: An open dataset for heterogeneous human activity recognition. Pattern Recognit. Lett. 2021;146:46–54. doi: 10.1016/j.patrec.2021.02.024. [DOI] [Google Scholar]
- 47.Gao X., Luo H., Wang Q., Zhao F., Ye L., Zhang Y. A human activity recognition algorithm based on stacking denoising autoencoder and lightGBM. Sensors (Switzerland) 2019;19:947. doi: 10.3390/s19040947. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Ignatov A. Real-time human activity recognition from accelerometer data using Convolutional Neural Networks. Appl. Soft Comput. 2018;62:915–922. doi: 10.1016/j.asoc.2017.09.027. [DOI] [Google Scholar]
- 49.Guan Y., Plötz T. Vol. 1. 2017. Ensembles of Deep LSTM Learners for Activity Recognition using Wearables; pp. 1–28. (Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies). [DOI] [Google Scholar]
- 50.Ordóñez F.J., Roggen D. Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition. Sensors. 2016;16:115. doi: 10.3390/s16010115. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Khaertdinov B., Ghaleb E., Asteriadis S. 2021 IEEE International Joint Conference on Biometrics (IJCB) IEEE; 2021. Contrastive Self-supervised Learning for Sensor-based Human Activity Recognition; pp. 1–8. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
-
•
All data generated or analyzed during this study are included in this published article.
-
•
This study does not report original code.
-
•
Any additional information required to reanalyze the data reported in this study is available from the lead contact upon request.




