Abstract
Automatic human activity recognition systems aim to capture the state of the user and its environment by exploiting heterogeneous sensors attached to the subject's body and permit continuous monitoring of numerous physiological signals reflecting the state of human actions. Successful identification of human activities can be immensely useful in healthcare applications for Ambient Assisted Living (AAL), for automatic and intelligent activity monitoring systems developed for elderly and disabled people. In this paper, we propose the method for activity recognition and subject identification based on random projections from high-dimensional feature space to low-dimensional projection space, where the classes are separated using the Jaccard distance between probability density functions of projected data. Two HAR domain tasks are considered: activity identification and subject identification. The experimental results using the proposed method with Human Activity Dataset (HAD) data are presented.
1. Introduction
The societies in the developed countries are rapidly aging. In 2006, almost 500 million people worldwide were 65 years of age or older. By 2030, that total number of aged people is projected to increase to 1 billion. The most rapid increase of aging population occurs in the developing countries, which will see a jump of 140% by 2030 [1]. Moreover, the world's population is expected to reach 9.3 billion by 2050 [2], and people who are above 60 years old will make up 28% of the population. Dealing with this situation will require huge financial resources to support the ever-increasing living cost, where human life expectancy is expected to reach 81 years by 2100.
As older people may have disorders of body functions or suffer from age-related diseases, the need for smart health assistance systems increases each year. A common method of monitoring geriatric patients is a physical observation, which is costly, requires a lot of human staff, and is increasingly infeasible in view of massive population aging in the following years. Many Ambient Assisted Living (AAL) applications such as care-providing robots, video surveillance systems, and assistive human-computer interaction technologies require human activity recognition. While the primary users of the AAL systems are of course the senior (elderly) people, the concept also applies to mentally and physically impaired people as well as people suffering from diabetes and obesity, who may need assistance at home, and people of any age interested in personal fitness monitoring. As a result, the sensor-based real-time monitoring system to support independent living at home has been a subject of many recent research studies in human activity recognition (HAR) domain [3–10].
Activity recognition can be defined as the process of how to interpret sensor data to classify a set of human activities [11]. HAR is a rapidly growing area of research that can provide valuable information on health, wellbeing, and fitness of monitored persons outside a hospital setting. Daily activity recognition using wearable technology plays a central role in the field of pervasive healthcare [12]. HAR has gained increased attention in the last decade due to the arrival of affordable and minimally invasive mobile sensing platforms such as smartphones. Smartphones are innovative platforms for HAR because of the availability of different wireless interfaces, unobtrusiveness, ease of use, high computing power and storage, and the availability of sensors, such as accelerometer, compass, and gyroscope, which meet the technical and practical hardware requirements for HAR tasks [13–15]. Moreover, technological development possibilities of other applications are still arising, including virtual reality systems. Therefore, these machines present a great possibility for the development of innovative technology dedicated for the AAL systems.
One of the key motivating factors for using mobile phone-based human activity recognition in the AAL systems is the relationship and correlation between the level of physical activity and the level of wellbeing of a person. Recording and analysing precise information on the person's activities are beneficial to keeping the progress and status of the disease (or mental condition) and can potentially improve the treatment of person's conditions and diseases, as well as decreasing the cost of care. Recognizing indoor and outdoor activities such as walking, running, or cycling can be useful to provide feedback to the caregiver about the patient's behaviour. When following the daily habits and routines of users, one can easily identify deviations from routines, which can assist the doctors in diagnosing conditions that would not be observed during routine medical examination. Another key enabler of the HAR technology is the possibility of providing independent living for the elderly as well as for patients with dementia and other mental pathologies, which could be monitored to prevent undesirable consequences of abnormal activities. Furthermore, by using persuasive techniques and gamification, HAR systems can be designed to interact with users to change their behaviour and lifestyles towards more active and healthier ones [16].
Recently, various intelligent systems based on mobile technologies have been constructed. HAR using smartphones or other types of portable or wearable sensor platforms has been used for assessing movement quality after stroke [17], such as upper extremity motion [18], for assessing gait characteristics of human locomotion for rehabilitation and diagnosis of medical conditions [19], for postoperative mobilization [20], for detecting Parkinson's disease, back pain, and hemiparesis [21], for cardiac rehabilitation [22], for physical therapy, for example, if a user is correctly doing the exercises recommended by a physician [23, 24], for detecting abnormal activities arising due to memory loss for dementia care [25, 26], for dealing with Alzheimer's [27] and neurodegenerative diseases such as epilepsy [28], for assessment of physical activity for children and adolescents suffering from hyperlipidaemia, hypertension, cardiovascular disease, and type 2 diabetes [29], for detecting falls [30, 31], for addressing physical inactivity when dealing with obesity [32], for analysing sleeping patterns [33], for estimating energy expenditures of a person to assess his/her healthy daily lifestyle [34], and for recognizing the user's intent in the domain of rehabilitation engineering such as smart walking support systems to assist motor-impaired persons and the elderly [35].
In this paper, we propose a new method for offline recognition of daily human activities based on feature dimensionality reduction using random projections [36] to low dimensionality feature space and using the Jaccard distance between kernel density probabilities as a decision function for classification of human activities.
The structure of the remaining parts of the paper is as follows. Section 2 presents the overview of related work in the smartphone-based HAR domain with a particular emphasis on the features extracted from the sensor data. Section 3 describes the proposed method. Section 4 evaluates and discusses the results. Finally, Section 5 presents the conclusions and discusses future work.
2. Overview of HAR Features and Related Work
All tasks of the HAR domain require correct identification of human activities from sensor data, which, in turn, requires that features derived from sensor data must be properly categorized and described. Next, we present an overview of features used in the HAR domain.
2.1. Features
While numerous features can be extracted from physical activity signals, increasing the number of features does not necessarily increase classification accuracy since the features may be redundant or may not be class-specific:
Time domain features (such as mean, median, variance, standard deviation, minimum, maximum, and root mean square, applied to the amplitude and time dimensions of a signal) are typically used in many practical HAR systems because of being less computationally intensive; thus, they can be easily extracted in real time.
Frequency-domain features require higher computational cost to distinguish between different human activities. Thus, they may not be suitable for real-time AAL applications.
Physical features are derived from a fundamental understanding of how a certain human movement would produce a specific sensor signal. Physical features are usually extracted from multiple sensor axes, based on the physical parameters of human movements.
Based on the extensive analysis of the literature and features used by other authors (esp. by Capela et al. [17], Mathie et al. [37], and Zhang and Sawchuk [38]), we have extracted 99 features of data, which are detailed in Table 1.
Table 1.
Catalogue of features.
Feature number | Description | Equation (notation) |
---|---|---|
4–6 | Acceleration (x-, y-, and z-axes) | a x, a y, a z |
| ||
7–9 | Gyroscope (x-, y-, and z-axes) | g x, g y, g z |
| ||
10–15 | Moving variance of 100 samples of acceleration and gyroscope data | , here x = a x, a y, a z, g x, g y, g z |
| ||
16-17 | Movement intensity of acceleration and gyroscope data |
|
| ||
18 | Movement intensity of difference between acceleration and gyroscope data | |
| ||
19–21 | Moving variance of 100 samples of movement intensity data | , here x = MIa, MIg, MIga |
| ||
22–24 | Polar coordinates of acceleration data |
φ
a = arctan(a
y, a
x), , z a = a z |
| ||
25–27 | Polar coordinates of gyroscope data |
φ
g = arctan(g
y, g
x), , z g = g z |
| ||
28–30 | Polar coordinates of difference between acceleration and gyroscope data |
φ
ag = arctan(a
y − g
y, a
x − g
x), , z ag = a z − g z |
| ||
31 | Simple moving average of acceleration data | |
| ||
32 | Simple moving average of gyroscope data | |
| ||
33 | Simple moving average of difference between acceleration and gyroscope data | |
| ||
34 | First eigenvalue of moving covariance between acceleration data | E a = eig1(cov(a x(1 : N), a y(1 : N), a z(1 : N))) |
| ||
35 | First eigenvalue of moving covariance between gyroscope data | E g = eig1(cov(g x(1 : N), g y(1 : N), g z(1 : N))) |
| ||
36 | First eigenvalue of moving covariance of difference between acceleration and gyroscope data | E ag = eig1(cov(a x − g x, a y − g y, a z − g z)) |
| ||
37–42 | Moving energy of acceleration and gyroscope data | , here x = a x, a y, a z, g x, g y, g z |
| ||
43–48 | Difference between moving maximum and moving minimum of acceleration and gyroscope data | , here x = a x, a y, a z, g x, g y, g z |
| ||
49 | Moving correlation between x- and y-axis of acceleration data | MCa xy = corr(a x, a y) |
| ||
50 | Moving correlation between x-axis and z-axis of acceleration data | MCa xz = corr(a x, a z) |
| ||
51 | Moving correlation between y-axis and z-axis of acceleration data | MCa yz = corr(a y, a z) |
| ||
52 | Moving correlation between x-axis and y-axis of gyroscope data | MCg xy = corr(g x, g y) |
| ||
53 | Moving correlation between x-axis and z-axis of gyroscope data | MCg xz = corr(g x, g z) |
| ||
54 | Moving correlation between y-axis and z-axis of gyroscope data | MCg yz = corr(g y, g z) |
| ||
55–57 | Projection of gyroscope data onto acceleration data | |
| ||
58 | Moving mean of orientation vector of acceleration data | , here |
| ||
59 | Moving variance of orientation vector of acceleration data | , |
| ||
60 | Moving energy of orientation vector of acceleration data | , here |
| ||
61–63 | Moving energy of difference between acceleration and gyroscope data | , here x = a x, a y, a z, y = g x, g y, g z |
| ||
64 | Moving energy of difference between x-axis and y-axis of acceleration data | |
| ||
65 | Moving energy of difference between x-axis and z-axis of acceleration data | |
| ||
66 | Moving energy of difference between y-axis and z-axis of acceleration data | |
| ||
67 | Moving mean of orientation vector of difference between acceleration and gyroscope data | , here |
| ||
68 | Moving variance of orientation vector of difference between acceleration and gyroscope data | , |
| ||
69 | Moving energy of orientation vector of difference between acceleration and gyroscope data | , here |
| ||
70 | Moving mean of orientation vector of gravity data | , here |
| ||
71 | Moving variance of orientation vector of gravity data | , |
| ||
72 | Moving energy of orientation vector of gravity data | , here |
| ||
73 | Moving mean of orientation vector of difference between acceleration and gravity data | , |
| ||
74 | Moving variance of orientation vector of difference between acceleration and gravity data |
, |
| ||
75 | Moving energy of orientation vector of difference between acceleration and gravity data | , |
| ||
76–81 | Moving cumulative sum of acceleration and gyroscope data | , here x = a x, a y, a z, g x, g y, g z |
| ||
82 | Simple moving average of cumulative sums of acceleration data | |
| ||
83 | Simple moving average of cumulative sums of gyroscope data | |
| ||
84 | Simple moving average of cumulative sums of difference between accelerometer and gyroscope data | |
| ||
85–90 | Moving 2nd-order cumulative sum of acceleration and gyroscope data | |
| ||
91–93 | Moving 2nd-order cumulative sum of differences between cumulative sums of acceleration and gyroscope data | |
| ||
94–96 | Polar coordinates of moving cumulative sum of acceleration data |
φ′a = arctan(MCSy
a, MCSx
a), , z′a = MCSz a |
| ||
97–99 | Polar coordinates of moving cumulative sum of gyroscope data |
φ′g = arctan(MCSy
g, MCSx
g), , z′g = MCSz g |
| ||
100–102 | Polar coordinates of moving cumulative sum of differences between acceleration and gyroscope data |
φ′ag = arctan(MCSy
a − MCSy
g, MCSx
a − MCSx
g), , z′ag = MCSz a − MCSz g |
2.2. Feature Selection
Feature selection is the process of selecting a subset of relevant features for use in construction of the classification model. Successful selection of features allows for simplification of models to make them easier to interpret, to decrease model training times, and to better understand difference between classes. Using feature selection allows removing redundant or irrelevant features without having an adverse effect on the classification accuracy. There are four basic steps in a typical feature selection method [58]: generation of candidate feature subset, an evaluation function for feature candidate subset, a generation stopping criterion, and a validation procedure.
Further, we analyse several feature selection methods used in the HAR domain.
ReliefF [59] is a commonly used filter method that ranks features by weighting them based on their relevance. Feature relevance is based on how well data instances are separated. For each data instance, the algorithm finds the nearest data point from the same class (hit) and nearest data points from different classes (misses).
Matlab's Rankfeatures ranks features by a given class separability criterion. Class separability measures include the absolute value of a statistic of a two-sample t-test, Kullback-Leibler distance, minimum attainable classification error, area between the empirical Receiver Operating Characteristic (ROC) curve and the random classifier slope, and the absolute value of the statistic of a two-sample unpaired Wilcoxon test. Measures are based on distributional characteristics of classes (e.g., mean, variance) for a feature.
Principal component analysis (PCA) is the simplest method to reduce data dimensionality. This reduced dimensional data can be used directly as features for classification. Given a set of N features, a PCA analysis will produce new data variables (PCA components) as linear combinations of the features with the highest variance in the subspace orthogonal to the preceding PCA component. As variability of the data can be captured by a relatively small number of PCs, PCA can achieve high level of dimensionality reduction. Several extensions of the PCA method are known such as kernel PCA, sparse PCA, and multilinear PCA.
Correlation-based Feature Selection (CFS) [60] is a filter algorithm that ranks subsets of features by a correlation-based heuristic evaluation function. A feature is considered to be a good one if it is relevant to the target concept but is not redundant to any of the other relevant features. Goodness of measure is expressed by a correlation between features, and CFS chooses the subset of features which has the highest measure. The chosen subset holds the property that features inside this subset have high correlation with the class and are unrelated to each other.
Table 2 summarizes the feature selection/dimensionality reduction methods in HAR.
Table 2.
Summary of feature selection/dimensionality reduction methods in HAR.
Method | Advantages | Disadvantages | Complexity |
---|---|---|---|
PCA | High dimensionality reduction; reduction of noise; lack of redundancy of data due to orthogonality of components | The covariance matrix is difficult to be evaluated accurately; even the simplest invariance could not be captured by the PCA unless the training data explicitly provides for it | O(p 2 n + p 3), where n are data points, each represented with p features |
| |||
ReliefF | Low computational complexity | Unstable due to random selection of instances | O(p · n · logn) |
| |||
Rankfeatures | Features highly correlated with already selected features are less likely to be included | It assumes that data classes are normally distributed | It depends upon class separability criterion |
| |||
CFS | It evaluates a subset of features rather than individual features | It fails to select locally predictive features when they are overshadowed by strong, globally predictive features |
A comprehensive review of feature selection algorithms in general as well as in the HAR domain can be found in [58, 61–63].
2.3. Summary
Related work in the HAR domain is summarized in Table 3. For each paper, the activities analysed, types of sensor data used, features extracted, classification method applied, and accuracy achieved (as given by the referenced papers) are given.
Table 3.
Summary of related works in the HAR domain.
Author | Activities | Sensor data | Features | Feature selection | Classification method | Accuracy |
---|---|---|---|---|---|---|
Atallah et al. [39] | Lying down, preparing food, eating and drinking, socialising, reading, dressing, walking, treadmill walking, vacuuming, wiping tables, running, treadmill, running, cycling, sitting down/getting up, and lying down/getting up | Acceleration sensors | Averaged entropy over 3 axes, main FFT frequency (averaged) over 3 axes, energy of the 0.2 Hz window centred around main frequency over total FFT energy (3-axis average), and averaged mean of cross covariance between every 2 axes | ReliefF, Simba, and MRMR | kNN, Bayesian classifier | 90% |
| ||||||
Bayat et al. [40] | Running, slow walk, fast walk, aerobic dancing, stairs up, and stairs down | Triaxial accelerometer | Mean along z-axis, MinMax, STD, and RMS for Am, APF along x-axis, y-axis, and z-axis, VarAPF, STD along x-axis, y-axis, and z-axis, RMS along x-axis, y-axis, and z-axis, correlation between z-axis and y-axis, and MinMax along x-axis, y-axis, and z-axis | Feature clustering | Multilayer perceptron, SVM, Random Forest, and Logit Boost | 81%–91% |
| ||||||
Berchtold et al. [41] | Standing, sitting, lying, walking, climbing stairs, cycling, and being stationary | Accelerometer | Variance, mean | None | Fuzzy inference | 97.3% |
| ||||||
Capela et al. [17] | Sitting, standing, and lying; ramp up and ramp down; stairs up and stairs down; transition between activities | Linear acceleration, gravity, and velocity sensors | Range, mean, standard deviation, kurtosis, moving average, covariance matrix, skewness, zero cross rate, and mean cross rate | None | Naïve-Bayes, Support Vector Machine, and j48 decision tree | 97% |
| ||||||
Gupta and Dallas [30] | Jumping, running, walking, sitting, sitting-to-standing, and standing-to-kneeling | Triaxial accelerometer | Energy, entropy, mean, variance, mean trend, windowed mean difference, variance trend, windowed variance difference, detrended fluctuation analysis coefficients, X-Z-energy, and max. difference acceleration | ReliefF, SFFS | kNN, Naive Bayes | 98% |
| ||||||
Henpraserttae et al. [42] | Sitting, lying, standing, and walking | Accelerometer | Mean and standard deviation | None | Rules and threshold based classification | 90% |
| ||||||
Hoque and Stankovic [43] | Leaving house, using toilet, taking shower, sleeping, preparing breakfast, preparing dinner, getting snack, getting drink, using washing machine, and using dishwasher | Location sensors (open/closed) | Magnitude | None | Custom clustering method | 64.5%–89.9% |
| ||||||
Iso and Yamazaki [44] | Walking, running, stairs up/down, and fast walking | Accelerometer | Wavelet components, periodograms, and information entropy | None | Bayesian probabilities | 80% |
| ||||||
Kose et al. [45] | Walking, running, biking, sitting, and standing | Accelerometer | Min., max., average, variance, FFT coefficients, and autocorrelation | None | Clustered kNN | 95.2%–97.5% |
| ||||||
Kwapisz et al. [46] | Walking, jogging, stairs up/down, sitting, and standing | Accelerometer | Mean, std. dev., average absolute difference, average resultant acceleration, time between peaks, and binned distribution | None | Decision tree, logistic regression, and MNN | 91.7% |
| ||||||
Lane et al. [47] | Driving, being stationary, running, and walking | GPS, accelerometer, and microphone | Mean, variance | None | Naïve-Bayes | 85–98% |
| ||||||
Lee and Cho [48] | Standing, walking, running, stairs up/down, shopping, and taking bus | Accelerometer | x-, y-, and z-axes acceleration values | None | Hierarchical HMM | 70%–90% |
| ||||||
Mannini and Sabatini [49] | Walking, walking carrying items, sitting & relaxing, working on computer, standing still, eating or drinking, watching TV, reading, running, bicycling, stretching, strength training, scrubbing, vacuuming, folding laundry, lying down and relaxing, brushing teeth, climbing stairs, riding elevator, and riding escalator | Acceleration sensors | DC component, energy, frequency-domain entropy, and correlation coefficients | SFFS (Pudil algorithm) | Continuous emissions, Hidden Markov Model | 99.1% |
| ||||||
Mathie et al. [37] | Various human movements, including resting, walking, and falling | Triaxial acceleration sensor | Integrated area under curve | None | Binary decision tree | 97.7% (sensitivity) 98.7% (specificity) |
| ||||||
Maurer et al. [50] | Walking, standing, sitting, running, and ascending and descending the stairs | Multiple sensors | Mean, root mean square, standard deviation, variance, mean absolute deviation, cumulative histogram, nth percentiles, interquartile range, zero crossing rate, mean crossing rate, and sq. length of X, Y | Correlation-based Feature Selection (CFS) | Decision trees (C4.5 algorithm), k-Nearest Neighbor, Naïve-Bayes, and Bayes Net | 80%–92% |
| ||||||
Miluzzo et al. [51] | Sitting, standing, walking, and running | Accelerometer, GPS, and audio | DFT, FFT features, mean, std. dev. and number of peaks per unit, and time deviation of DFT power | None | Decision tree | 79% |
| ||||||
Pärkkä et al. [52] | Lying down, rowing, ex-biking, sitting/standing, running, and Nordic walking | GPS, audio, altitude, EKG, accelerometer, compass, humidity, light, temperature, heart rate, pulse, respiratory effort, and skin resistance | Peak frequency of up-down chest acceleration, median of up-down chest acceleration, peak power of up-down chest acceleration, variance of back-forth chest acceleration, sum of variances of 3D wrist acceleration, and power ratio of frequency bands 1–1.5 Hz and 0.2–5 Hz measured from chest magnetometer | Heuristic | Decision tree | 86% |
| ||||||
Saponas et al. [53] | Walking, jogging | Accelerometer | 124 features: Nike + iPod Packet Payload, magnitude (mean, std. dev., min., max., and min. minus max.), frequency (energy in each of the first 10 frequency components of DFT, energy in each band of 10 frequency components, largest frequency component, and index of the largest frequency component) | None | Naïve-Bayesian Network | 97.4% (within-person), 99.48% (cross-person) |
| ||||||
Siirtola and Röning [54] | Walking, running, cycling, driving, sitting, and standing | Accelerometer | Magnitude, std., mean, min., max., percentiles (10, 25, 50, 75, and 90), and sum and square sum of observations above/below percentile (5, 10, 25, 75, 90, and 95) of magnitude acceleration and square sum of x & z | None | Decision tree + kNN/QDA | 95% |
| ||||||
Sohn et al. [55] | Walking, driving, and dwelling | GPS | Spearman rank correlation, variance, and mean Euclidean distance over a window of measurements | None | Logistic regression | 85% |
| ||||||
Yang [56] | Sitting, Standing, walking, running, driving, and bicycling | Accelerometer | Mean, std., zero crossing rate, 75th percentile, interquartile, spectrum centroid, entropy, and cross-correlation | None | Decision tree, Naïve-Bayes, kNN, and SVM | 90% |
| ||||||
Zhu and Sheng [57] | Sitting, standing, lying, walking, sitting-to-standing, standing-to-sitting, lying-to-sitting, and sitting-to-lying | 3D acceleration | Mean, variance | None | Neural network ensemble | 67%–98% |
3. Method
3.1. General Scheme
The typical steps for activity recognition are preprocessing, segmentation, feature extraction, dimensionality reduction (feature selection), and classification [24]. The main steps of activity recognition include (a) preprocessing of sensor data (e.g., denoising), (b) feature extraction, (c) dimension reduction, and (d) classification. The preprocessing step includes noise removal and representation of raw data. The feature extraction step is used to reduce large input sensor data to a smaller set of features (feature vector), which preserves information contained in the original data. The dimensionality reduction step can be applied to remove the irrelevant (or less relevant) features and reduce the computational complexity and increase the performance of the activity recognition process. The classification step is used to map the feature set to a set of activities.
In this paper, we do not focus on data preprocessing and feature extraction but rather on dimensionality reduction and classification steps, since these two are crucial for further efficiency of AAL systems. The proposed method for human activity recognition is based on feature dimensionality reduction using random projections [36] and classification using kernel density function estimate as a decision function (Figure 1).
Figure 1.
General scheme of the proposed method.
3.2. Description of the Method
During random projection, the original d-dimensional data is projected to a k-dimensional (k ≪ d) subspace using a random k × d matrix R. The projection of the data onto a lower k-dimensional subspace is X k×N RP = R k×d X d×N, where X d×N is the original set of N d-dimensional observations. In the derived projection, the distances between the points are approximately preserved, if points in a vector space are projected onto a randomly selected subspace of suitably high dimension (see the Johnson-Lindenstrauss lemma [64]). The random matrix R is selected as proposed by Achlioptas [36] as follows:
(1) |
Given the low dimensionality of the target space, we can treat the projection of low-dimensional observations onto each dimension as a set of random variables for which the probability density function (PDF) can be estimated using kernel density estimation (KDE) (or Parzen window) method [65].
If x 1, x 2,…, x N is a sample of a random variable, then the kernel density approximation of its probability density function is
(2) |
where K is some kernel and h is the bandwidth (smoothing parameter). K is taken to be a standard Gaussian function with mean zero and variance 1 of the examined data features:
(3) |
For a two-dimensional case, the bivariate probability density function is calculated as a product of univariate probability functions as follows:
(4) |
Here, x and y are data in each dimension, respectively.
However, each random projection produces a different mapping of the original data points which reveals only a part of the data manifold in higher-dimensional space. In case of the binary classification problem, we are interested in a mapping that separates data points belonging to two different classes best.
As a criterion for estimating the mapping, we use the Jaccard distance metric between two probability density estimates of data points representing each class. The advantage of the Jaccard distance metric as compared to other metrics of distance such as Kullback-Leibler (KL) divergence and Hellinger distance is its adaptability to multidimensional spaces where compared points show relations to different subsets. Therefore, it is well adapted to the developed model of human activity features, where according to description in the previous section we have divided them into some sets of actions. Furthermore, the computational complexity of the Hellinger distance is very high, while KL divergence might be unbounded.
The Jaccard distance, which measures dissimilarity between sample sets, is obtained by subtracting the Jaccard coefficient from 1 or, equivalently, by dividing the difference of the sizes of the union and the intersection of two sets by the size of the union:
(5) |
In the proposed model, the best random projection with the smallest overlapping area is selected (see an example in Figure 2).
Figure 2.
Graphical illustration of good separation versus bad separation of kernel density estimation functions (Subject 1, Trial 1, Walking Forward versus Walking Upstairs; 2nd dimension).
To explore the performance and correlation among features visually, a series of scatter plots in a 2D feature space is shown in Figure 3. The horizontal and vertical axes represent two different features. The points in different colours represent different human activities.
Figure 3.
Example of classification: walking versus running (Subject 1, Trial 1) classes randomly projected in a bidimensional feature subspace.
In case of multiple classes, the method works as a one-class classifier: recognizing instances of a positive class, while all instances of other classes are recognized as outliers of the positive class.
3.3. Algorithm
The pseudocode of the algorithms for finding the best projection and using it for classification in low-dimensional space is presented in Pseudocodes 1 and 2, respectively.
Pseudocode 1.
Pseudocode of FindBestProjection.
Pseudocode 2.
Pseudocode of binary classification.
4. Experiments
4.1. Dataset
To evaluate the performance of the proposed approach for HAR from the smartphone data, we used the part of the dataset (USC Human Activity Dataset [38]) recorded using the MotionNode device (sampling rate: 100 Hz; 3-axis accelerometer range: ±6 g; 3-axis gyroscope range: ±500 dps). The dataset consists of records recorded with 14 subjects (7 male, 7 female; age: 21–49) of 12 activities, 5 trials each. During data acquisition, MotionNode was attached on the front right hip of subjects.
The recorded low-level activities are as follows: Walking Forward (WF), Walking Left (WL), Walking Right (WR), Walking Upstairs (WU), Walking Downstairs (WD), Running Forward (RF), Jumping Up (JU), Sitting (Si), Standing (St), Sleeping (Sl), Elevator Up (EU), and Elevator Down (ED). Each record consists of the following attributes: date, subject number, age, height, weight, activity name, activity number, trial number, sensor location, orientation, and readings. Sensor readings consist of 6 readings: acceleration along x-, y-, and z-axes and gyroscope along x-, y-, and z-axes. Each trial was performed on different days at various indoor and outdoor locations.
4.2. Results
In Table 4, we describe the top three best features from Table 1 (see column Feature number) ranked by the Matlab Rankfeatures function using the entropy criterion.
Table 4.
Top features for binary classification of human activities.
Activity | WL | WR | WU | WD | RF | JU | Si | St | Sl | EU | ED |
---|---|---|---|---|---|---|---|---|---|---|---|
WF | 97, 88, 91 | 85, 100, 88 | 15, 63, 42 | 14, 39, 34 | 10, 15, 60 | 10, 19, 37 | 17, 18, 30 | 19, 36, 35 | 34, 19, 36 | 15, 36, 35 | 15, 42, 63 |
WL | 97, 91, 88 | 97, 63, 42 | 87, 97, 86 | 10, 37, 19 | 10, 37, 19 | 17, 18, 30 | 19, 36, 34 | 34, 19, 36 | 15, 36, 35 | 15, 42, 63 | |
WR | 63, 42, 15 | 34, 87, 39 | 10, 59, 37 | 10, 37, 19 | 17, 18, 30 | 19, 34, 14 | 34, 19, 85 | 15, 36, 35 | 15, 42, 63 | ||
WU | 87, 39, 78 | 63, 42, 15 | 10, 19, 63 | 17, 18, 26 | 19, 14, 62 | 34, 19, 35 | 15, 36, 35 | 15, 42, 63 | |||
WD | 65, 38, 34 | 10, 19, 60 | 17, 18, 7 | 19, 10, 20 | 19, 34, 35 | 15, 36, 35 | 15, 42, 63 | ||||
RF | 35, 36, 62 | 38, 36, 17 | 19, 10, 15 | 34, 19, 35 | 15, 36, 35 | 15, 42, 63 | |||||
JU | 4, 22, 16 | 19, 10, 15 | 19, 34, 36 | 15, 42, 63 | 15, 42, 63 | ||||||
Si | 22, 5, 38 | 22, 4, 85 | 15, 42, 63 | 59, 60, 15 | |||||||
St | 76, 85, 39 | 10, 15, 36 | 59, 11, 60 | ||||||||
Sl | 15, 22, 42 | 59, 60, 94 | |||||||||
EU | 59, 60, 10 |
WF: Walking Forward; WL: Walking Left; WR: Walking Right; WU: Walking Upstairs; WD: Walking Downstairs; RF: Running Forward; JU: Jumping Up; Si: Sitting; St: Standing; Sl: Sleeping; EU: Elevator Up; ED: Elevator Down.
The results of feature ranking presented in Table 5 can be summarized as follows:
-
(i)
For Walking Forward, Walking Left, and Walking Right, the important features are moving variance of acceleration and gyroscope data, movement intensity of gyroscope data, moving variance of movement intensity of acceleration data, first eigenvalue of moving covariance between acceleration data, and polar angle of moving cumulative sum of gyroscope data.
-
(ii)
For Walking Upstairs and Walking Downstairs, moving variance of gyroscope along z-axis, movement intensity of gyroscope data, and moving variance of movement intensity are the most important.
-
(iii)
For Running Forward, moving variance of 100 samples of acceleration along x-axis, moving variance of 100 samples of gyroscope along z-axis, and moving energy of acceleration are distinguishing features.
-
(iv)
For Jumping Up, the most important features are moving variance of acceleration, moving variance of movement intensity, and moving energy of acceleration.
-
(v)
For Sitting, movement intensity of gyroscope data and movement intensity of difference between acceleration and gyroscope data are the most important.
-
(vi)
For Standing, moving variance of movement intensity of acceleration data, moving variance of acceleration along x-axis, and first eigenvalue of moving covariance of difference between acceleration and gyroscope data are the most distinctive.
-
(vii)
For Sleeping, the most prominent features are first eigenvalue of moving covariance between acceleration data and moving variance of movement intensity of acceleration data.
-
(viii)
For Elevator Up and Elevator Down, the most commonly selected feature is moving variance of z-axis of gyroscope data. Other prominent features are first eigenvalue of moving covariance of difference between acceleration and gyroscope data and moving energy of z-axis of gyroscope data.
Table 5.
The confusion matrix of within-subject activity classification using Rankfeatures.
Activity | WF | WL | WR | WU | WD | RF | JU | Si | St | Sl | EU | ED |
---|---|---|---|---|---|---|---|---|---|---|---|---|
WF | 1 | 0.774 | 0.980 | 0.874 | 0.985 | 0.996 | 0.980 | 0.997 | 0.999 | 1 | 0.999 | 1 |
WL | 0.774 | 1 | 0.989 | 0.968 | 0.958 | 0.998 | 0.996 | 0.951 | 0.999 | 1 | 0.999 | 1 |
WR | 0.980 | 0.989 | 1 | 0.798 | 0.998 | 0.997 | 0.988 | 0.988 | 0.971 | 1 | 0.981 | 1 |
WU | 0.874 | 0.968 | 0.798 | 1 | 0.708 | 0.985 | 0.979 | 0.962 | 0.998 | 1 | 0.971 | 1 |
WD | 0.985 | 0.958 | 0.998 | 0.708 | 1 | 0.992 | 0.878 | 0.967 | 0.850 | 1 | 0.986 | 1 |
RF | 0.996 | 0.998 | 0.997 | 0.985 | 0.992 | 1 | 0.978 | 0.991 | 0.996 | 0.957 | 0.994 | 1 |
JU | 0.980 | 0.996 | 0.988 | 0.979 | 0.878 | 0.978 | 1 | 0.973 | 1 | 0.929 | 1 | 1 |
Si | 0.997 | 0.951 | 0.988 | 0.962 | 0.967 | 0.991 | 0.973 | 1 | 0.987 | 1 | 0.126 | 0.992 |
St | 0.999 | 0.999 | 0.971 | 0.999 | 0.850 | 0.995 | 1 | 0.987 | 1 | 1 | 0.326 | 0.887 |
Sl | 1 | 1 | 1 | 1 | 1 | 0.957 | 0.930 | 1 | 1 | 1 | 1 | 0.992 |
EU | 0.999 | 0.999 | 0.981 | 0.971 | 0.986 | 0.994 | 1 | 0.126 | 0.326 | 1 | 1 | 0.697 |
ED | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 0.992 | 0.887 | 0.992 | 0.697 | 1 |
Mean | 0.965 | 0.969 | 0.974 | 0.937 | 0.944 | 0.990 | 0.975 | 0.911 | 0.918 | 0.989 | 0.840 | 0.964 |
Baseline | 0.650 | 0.616 | 0.616 | 0.712 | 0.713 | 0.621 | 0.641 | 0.627 | 0.642 | 0.651 | 0.640 | 0.628 |
WF: Walking Forward; WL: Walking Left; WR: Walking Right; WU: Walking Upstairs; WD: Walking Downstairs; RF: Running Forward; JU: Jumping Up; Si: Sitting; St: Standing; Sl: Sleeping; EU: Elevator Up; ED: Elevator Down.
These results can be considered as consistent from what can be expected from the physical analysis of human motions in the analysed dataset.
The evaluation of HAR classification algorithms is usually made through the statistical analysis of the models using the available experimental data. The most common method is the confusion matrix which allows representing the algorithm performance by clearly identifying the types of errors (false positives and negatives) and correctly predicted samples over the test data.
The confusion matrix for within-subject activity recognition using Matlab's Rankfeatures is detailed in Table 5. The classification was performed using 5-fold cross-validation, using 80% of data for training and 20% of data for testing. Grand mean accuracy is 0.9552; grand mean precision is 0.9670; grand mean sensitivity is 0.9482; grand mean specificity is 0.9569; grand mean recall is 0.9482; grand mean F-score is 0.9482. The baseline accuracy was calculated using only the top 2 features selected by Rankfeatures, but without using random projections. The results show that features derived using random projections are significantly better than features derived using a common feature selection algorithm.
To take a closer look at the classification result, Table 5 shows the confusion table for classification of activities. The overall averaged recognition accuracy across all activities is 95.52%, with 11 out of 12 activities having accuracy values higher than 90%. If we examine the recognition performance for each activity individually, Running Forward, Jumping Up, and Sleeping will have very high accuracy values. For Running Forward, the accuracy of 99.0% is achieved. Interestingly, the lowest accuracy was achieved for Elevator Up activity, only 84.0%, while it was most often misclassified with Sitting and Standing. Elevator Down is misclassified with Elevator Up (only 69.7% accuracy). This result makes sense since Sitting on a chair, Standing, and Standing in a moving elevator are static activities, and we expect difficulty in differentiating different static activities. Also, there is some misclassification when deciding on a specific direction of activity; for example, Walking Left is confused with Walking Forward (77.4% accuracy) and Walking Upstairs (87.4% accuracy). Walking Upstairs is also confused with Walking Right (79.8% accuracy) and Walking Downstairs (70.8% accuracy). This is due to the similarity of any walk-related activities.
For comparison, the confusion matrix for within-subject activity recognition obtained using the proposed method with ReliefF feature selection is detailed in Table 6. The classification was performed using 5-fold cross-validation, using 80% of data for training and 20% of data for testing. Grand mean accuracy is 0.932; grand mean precision is 0.944; grand mean sensitivity is 0.939; grand mean specificity is 0.933; grand mean recall is 0.939; grand mean F-score is 0.922.
Table 6.
The confusion matrix of within-subject activity classification using ReliefF.
Activity | WF | WL | WR | WU | WD | RF | JU | Si | St | Sl | EU | ED |
---|---|---|---|---|---|---|---|---|---|---|---|---|
WF | 1.000 | 0.998 | 0.892 | 0.931 | 0.993 | 1.000 | 0.999 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 |
WL | 0.998 | 1.000 | 0.853 | 0.989 | 0.997 | 1.000 | 1.000 | 0.999 | 1.000 | 0.896 | 0.971 | 1.000 |
WR | 0.892 | 0.853 | 1.000 | 0.789 | 0.964 | 1.000 | 1.000 | 0.999 | 1.000 | 0.853 | 1.000 | 1.000 |
WU | 0.931 | 0.989 | 0.789 | 1.000 | 0.702 | 1.000 | 1.000 | 0.996 | 0.956 | 0.992 | 0.999 | 1.000 |
WD | 0.993 | 0.997 | 0.964 | 0.702 | 1.000 | 0.965 | 0.998 | 0.655 | 0.997 | 0.982 | 1.000 | 0.975 |
RF | 1.000 | 1.000 | 1.000 | 1.000 | 0.965 | 1.000 | 0.688 | 0.999 | 1.000 | 1.000 | 1.000 | 1.000 |
JU | 0.999 | 1.000 | 1.000 | 1.000 | 0.998 | 0.688 | 1.000 | 1.000 | 1.000 | 0.993 | 1.000 | 1.000 |
Si | 1.000 | 0.999 | 0.999 | 0.996 | 0.655 | 0.999 | 1.000 | 1.000 | 0.491 | 0.967 | 0.328 | 0.313 |
St | 1.000 | 1.000 | 1.000 | 0.956 | 0.997 | 1.000 | 1.000 | 0.491 | 1.000 | 0.766 | 0.528 | 0.901 |
Sl | 1.000 | 0.896 | 0.853 | 0.992 | 0.982 | 1.000 | 0.993 | 0.967 | 0.766 | 1.000 | 1.000 | 1.000 |
EU | 1.000 | 0.971 | 1.000 | 0.999 | 1.000 | 1.000 | 1.000 | 0.328 | 0.528 | 1.000 | 1.000 | 0.765 |
ED | 1.000 | 1.000 | 1.000 | 1.000 | 0.975 | 1.000 | 1.000 | 0.313 | 0.901 | 1.000 | 0.765 | 1.000 |
Mean | 0.984 | 0.975 | 0.946 | 0.946 | 0.936 | 0.971 | 0.973 | 0.812 | 0.887 | 0.954 | 0.883 | 0.913 |
Baseline | 0.621 | 0.637 | 0.600 | 0.695 | 0.703 | 0.644 | 0.618 | 0.628 | 0.640 | 0.644 | 0.635 | 0.642 |
WF: Walking Forward; WL: Walking Left; WR: Walking Right; WU: Walking Upstairs; WD: Walking Downstairs; RF: Running Forward; JU: Jumping Up; Si: Sitting; St: Standing; Sl: Sleeping; EU: Elevator Up; ED: Elevator Down.
The baseline accuracy was calculated using only the top 2 features selected using ReliefF, but without using random projections. Again, the results show that features derived using random projections are significantly better than features derived using the ReliefF method only.
Surprisingly, though the classification accuracy of the specific activities differed, the mean accuracy metric results are quite similar (but still worse, if grand mean values are considered). The features identified using ReliefF feature selection were better at separating Walking Forward from Walking Left and Standing from Elevator Up activities but proved worse for separating other activities such as Sitting from Standing.
For subject identification, the data from all physical actions is used to train the classifier. Here, we consider one-versus-all subject identification problem. Therefore, the data of one subject is defined as positive class, and the data of all other subjects is defined as negative class. In this case, also 5-fold cross-validation was performed, using 80% of data for training and 20% of data for testing. The results of one-versus-all subject identification using all activities for training and testing are presented in Table 7. While the results are not very good, they still are better than random baselines: grand mean accuracy is 0.477; precision is 0.125; recall is 0.832; and F-score is 0.210.
Table 7.
Results of one-versus-all subject identification (all activities).
Subjects | Accuracy | Precision | Recall | F-score |
---|---|---|---|---|
S1 | 0.657 | 0.142 | 0.744 | 0.239 |
S2 | 0.273 | 0.075 | 0.917 | 0.139 |
S3 | 0.549 | 0.078 | 0.716 | 0.140 |
S4 | 0.496 | 0.073 | 0.697 | 0.132 |
S5 | 0.323 | 0.068 | 0.931 | 0.127 |
S6 | 0.863 | 0.220 | 0.637 | 0.328 |
S7 | 0.265 | 0.055 | 0.920 | 0.103 |
S8 | 0.107 | 0.071 | 0.985 | 0.132 |
S9 | 0.683 | 0.229 | 0.967 | 0.370 |
S10 | 0.156 | 0.091 | 0.943 | 0.166 |
S11 | 0.755 | 0.263 | 0.905 | 0.407 |
S12 | 0.689 | 0.167 | 0.533 | 0.254 |
S13 | 0.373 | 0.115 | 0.881 | 0.203 |
S14 | 0.493 | 0.112 | 0.866 | 0.198 |
Mean | 0.477 | 0.123 | 0.832 | 0.210 |
Random baseline | 0.071 | 0.071 | 0.929 | 0.133 |
If an activity of a subject has been established, separate classifiers for each activity can be used for subject identification. In this case, also 5-fold cross-validation was performed, using 80% of data for training and 20% of data for testing, and the results are presented in Table 8. The grand mean accuracy is 0.720, which is better than random baseline. However, if we consider only the top three walking-related activities (Walking Forward, Walking Left, or Walking Right), the mean accuracy is 0.944.
Table 8.
Results of one-versus-all subject identification for specific activities.
Activity | Accuracy | F |
---|---|---|
Walking Forward | 0.947 | 0.727 |
Walking Left | 0.955 | 0.769 |
Walking Right | 0.931 | 0.722 |
Walking Upstairs | 0.857 | 0.551 |
Walking Downstairs | 0.833 | 0.497 |
Running Forward | 0.832 | 0.496 |
Jumping Up | 0.814 | 0.453 |
Sitting | 0.506 | 0.391 |
Standing | 0.722 | 0.432 |
Sleeping | 0.589 | 0.292 |
Elevator Up | 0.337 | 0.235 |
Elevator Down | 0.318 | 0.232 |
Mean | 0.720 | 0.483 |
Random baseline | 0.071 | 0.133 |
Finally, we can simplify the classification problem to binary classification (i.e., recognize one subject against another). This simplification can be motivated by the assumption that only a few people are living in an AAL home (far less than 14 subjects in the analysed dataset). Then, the data from a pair of subjects performing a specific activity is used for classification and training. Separate classifiers are built for each pair of subjects, the results are evaluated using 5-fold cross-validation, and the results are averaged. The results are presented in Table 9. Note that the grand mean accuracy has increased to 0.947, while, for the top three walking-related activities (Walking Forward, Walking Left, or Walking Right), the grand mean accuracy is 0.992.
Table 9.
Accuracy of binary subject identification using separate activities.
Activity | Accuracy | F |
---|---|---|
Walking Forward | 0.992 | 0.987 |
Walking Left | 0.989 | 0.987 |
Walking Right | 0.993 | 0.993 |
Walking Upstairs | 0.977 | 0.970 |
Walking Downstairs | 0.974 | 0.971 |
Running Forward | 0.980 | 0.974 |
Jumping Up | 0.983 | 0.980 |
Sitting | 0.883 | 0.859 |
Standing | 0.940 | 0.932 |
Sleeping | 0.956 | 0.953 |
Elevator Up | 0.856 | 0.847 |
Elevator Down | 0.846 | 0.822 |
Mean | 0.947 | 0.939 |
Random baseline | 0.5 | 0.5 |
5. Evaluation and Discussion
Random projections have been used in the HAR domain for data dimensionality reduction in activity recognition from noisy videos [69], feature compression for head pose estimation [70], and feature selection for activity motif discovery [71]. The advantages of random projections are the simplicity of their implementation and their scalability, robustness to noise, and low computational complexity: constructing the random matrix R and projecting the d × N data matrix into k dimensions are of order O(dkN).
The HAD dataset has been used in HAR research by other authors, too. Using the same HAD dataset, Zheng [66] has achieved 95.6% accuracy. He used the means and variances of magnitude and angles as the activity features and the magnitude and angles that were produced by a triaxial acceleration vector. Classifier used the Least Squares Support Vector Machine (LS-SVM) and Naïve-Bayes (NB) algorithm to distinguish different activity classes. Sivakumar [67] achieved 84.3% overall accuracy using symbolic approximation of time series of accelerometer and gyroscope signal. Vaka [68] achieved 90.7% accuracy for within-person classification and 88.6% accuracy for interperson classification using Random Forest. The features used for the recognition were time domain features: mean, standard deviation, correlation between X and Y, correlation between Y and Z, correlation between X and Z, and root mean square of a signal. Our results (95.52% accuracy), obtained using the proposed method, are very similar to the best results of Zheng for activity recognition task.
The results obtained by different authors using the USC-HAD dataset are summarized in Table 10.
Table 10.
Summary of HAR results using USC-HAD dataset.
Reference | Features | Classification method | Accuracy |
---|---|---|---|
Zheng [66] | Means and variances of magnitude and angles of acceleration along x-, y- & z-axes | Least Squares Support Vector Machine (LS-SVM), Naïve-Bayes (NB) | 95.6% |
Sivakumar [67] | Accelerometer and gyroscope data | Symbolic approximation | 84.3% |
Vaka [68] | Mean, std. dev., correlation between X & Y, Y & Z, and X & Z, and RMS | Random Forest | 90.7% |
This paper | 99 times, frequency and physical features | Heuristic (random projections + PDFs + Jaccard distance) | 95.52% |
We think that it would be difficult to achieve even higher results due to some problems with the analysed dataset, which include a set of problems inherent to many Human Activity Datasets as follows:
Accurate Labelling of All Activities. Existing activity recognition algorithms usually are based on supervised learning where the training data depends upon accurate labelling of all human activities. Collecting consistent and reliable data is a very difficult task since some activities may have been marked by users with wrong labels.
Transitionary/Overlapping Activities. Often people do several activities at the same time. The transition states (such as walking-standing, lying-standing) can be treated as additional states, and the recognition model can be trained with respect to these states to increase the accuracy.
Context Problem. It occurs when the sensors are placed at an inappropriate position relative to the activity being measured. For example, with accelerometer-based HAR, the location where the phone is carried, such as in the pocket or in the bag, impacts the classification performance.
Subject Sensitivity. It measures dependency of the trained classification model upon the specifics of user.
Weak Link between Basic Activities and More Complex Activities. For example, it is rather straightforward to detect whether the user is running, but inferring whether the user is running away from danger or jogging in a park is different.
Spurious Data. Most published studies handle the problem of the fuzzy borders by manual data cropping.
6. Conclusion
Monitoring and recognizing human activities are important for assessing changes in physical and behavioural profiles of the population over time, particularly for the elderly and impaired and patients with chronic diseases. Although a wide variety of sensors are being used in various devices for activity monitoring, the positioning of the sensors, the selection of relevant features for different activity groups, and providing context to sensor measurements still pose significant research challenges.
In this paper, we have reviewed the stages needed to implement a human activity recognition method for automatic classification of human physical activity from on-body sensors. A major contribution of the paper lies in pursuing the random projections based approach for feature dimensionality reduction. The results of extensive testing performed on the USC-HAD dataset (we have achieved overall accuracy of within-person classification of 95.52% and interperson identification accuracy of 94.75%) reveal the advantages of the proposed approach. Gait-related activities (Walking Forward, Walking Left, and Walking Right) allowed the best identification of subjects opening the way for a multitude of applications in the area of gait-based identification and verification.
Future work will concern the validation of the proposed method using other datasets of human activity data as well as integration of the proposed method in the wearable sensor system we are currently developing for applications in indoor human monitoring.
Acknowledgments
The authors would like to acknowledge the contribution of the COST Action IC1303 AAPELE: Architectures, Algorithms and Platforms for Enhanced Living Environments.
Competing Interests
The authors declare that there are no competing interests regarding the publication of this paper.
References
- 1.U. S. State Department. Why Population Aging Matters: A Global Perspective. 2007. [Google Scholar]
- 2.Department of Economic and Social Affairs. World Population to 2300. New York, NY, USA: United Nations; 2004. [Google Scholar]
- 3.Turaga P., Chellappa R., Subrahmanian V. S., Udrea O. Machine recognition of human activities: a survey. IEEE Transactions on Circuits and Systems for Video Technology. 2008;18(11):1473–1488. doi: 10.1109/tcsvt.2008.2005594. [DOI] [Google Scholar]
- 4.Poppe R. A survey on vision-based human action recognition. Image and Vision Computing. 2010;28(6):976–990. doi: 10.1016/j.imavis.2009.11.014. [DOI] [Google Scholar]
- 5.Aggarwal J. K., Ryoo M. S. Human activity analysis: a review. ACM Computing Surveys. 2011;43(3, article 16) doi: 10.1145/1922649.1922653. [DOI] [Google Scholar]
- 6.Ke S.-R., Thuc H. L. U., Lee Y.-J., Hwang J.-N., Yoo J.-H., Choi K.-H. A review on video-based human activity recognition. Computers. 2013;2(2):88–131. doi: 10.3390/computers2020088. [DOI] [Google Scholar]
- 7.Ann O. C. Human activity recognition: a review. Proceedings of the of IEEE International Conference on Control System, Computing and Engineering (ICCSCE '14); November 2014; Batu Ferringhi, Malaysia. pp. 389–393. [DOI] [Google Scholar]
- 8.Aggarwal J. K., Xia L. Human activity recognition from 3D data: a review. Pattern Recognition Letters. 2014;48:70–80. doi: 10.1016/j.patrec.2014.04.011. [DOI] [Google Scholar]
- 9.Vrigkas M., Nikou C., Kakadiaris I. A. A review of human activity recognition methods. Frontiers in Robotics and AI. 2015;2, article 28 doi: 10.3389/frobt.2015.00028. [DOI] [Google Scholar]
- 10.Ziaeefard M., Bergevin R. Semantic human activity recognition: a literature review. Pattern Recognition. 2015;48(8):2329–2345. doi: 10.1016/j.patcog.2015.03.006. [DOI] [Google Scholar]
- 11.Incel O. D., Kose M., Ersoy C. A review and taxonomy of activity recognition on mobile phones. BioNanoScience. 2013;3(2):145–171. doi: 10.1007/s12668-013-0088-3. [DOI] [Google Scholar]
- 12.Osmani V., Balasubramaniam S., Botvich D. Human activity recognition in pervasive health-care: supporting efficient remote collaboration. Journal of Network and Computer Applications. 2008;31(4):628–655. doi: 10.1016/j.jnca.2007.11.002. [DOI] [Google Scholar]
- 13.Lara Ó. D., Labrador M. A. A survey on human activity recognition using wearable sensors. IEEE Communications Surveys and Tutorials. 2013;15(3):1192–1209. doi: 10.1109/SURV.2012.110112.00192. [DOI] [Google Scholar]
- 14.Shoaib M., Bosch S., Incel O. D., Scholten H., Havinga P. J. M. A survey of online activity recognition using mobile phones. Sensors. 2015;15(1):2059–2085. doi: 10.3390/s150102059. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Ortiz J. L. R. Smartphone-Based Human Activity Recognition. Springer Theses; 2015. [Google Scholar]
- 16.Purpura S., Schwanda V., Williams K., Stubler W., Sengers P. Fit4Life: The design of a persuasive technology promoting healthy behavior and ideal weight. Proceedings of the 29th Annual CHI Conference on Human Factors in Computing Systems (CHI '11); May 2011; New York, NY, USA. ACM; pp. 423–432. [DOI] [Google Scholar]
- 17.Capela N. A., Lemaire E. D., Baddour N. Feature selection for wearable smartphone-based human activity recognition with able bodied, elderly, and stroke patients. PLoS ONE. 2015;10(4) doi: 10.1371/journal.pone.0124414.e0124414 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Patel S., Hughes R., Hester T., et al. A novel approach to monitor rehabilitation outcomes in stroke survivors using wearable technology. Proceedings of the IEEE. 2010;98(3):450–461. doi: 10.1109/JPROC.2009.2038727. [DOI] [Google Scholar]
- 19.Tao W., Liu T., Zheng R., Feng H. Gait analysis using wearable sensors. Sensors. 2012;12(2):2255–2283. doi: 10.3390/s120202255. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Appelboom G., Taylor B. E., Bruce E., et al. Mobile phone-connected wearable motion sensors to assess postoperative mobilization. JMIR mHealth and uHealth. 2015;3(3, article e78) doi: 10.2196/mhealth.3785. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Cheung V. H., Gray L., Karunanithi M. Review of accelerometry for determining daily activity among elderly patients. Archives of Physical Medicine and Rehabilitation. 2011;92(6):998–1014. doi: 10.1016/j.apmr.2010.12.040. [DOI] [PubMed] [Google Scholar]
- 22.Bidargaddi N., Sarela A., Klingbeil L., Karunanithi M. Detecting walking activity in cardiac rehabilitation by using accelerometer. Proceedings of the International Conference on Intelligent Sensors, Sensor Networks and Information (ISSNIP '07); December 2007; pp. 555–560. [DOI] [Google Scholar]
- 23.Bao L., Intille S. S. Activity recognition from user-annotated acceleration data. In: Ferscha A., Mattern F., editors. Pervasive Computing. Vol. 3001. Berlin, Germany: Springer; 2004. pp. 1–17. (Lecture Notes in Computer Science). [DOI] [Google Scholar]
- 24.Avci A., Bosch S., Marin-Perianu M., Marin-Perianu R., Havinga P. Activity recognition using inertial sensing for healthcare, wellbeing and sports applications: a survey. Proceedings of the 23rd International Conference on Architecture of Computing Systems (ARCS 10); February 2010; Hannover, Germany. pp. 1–10. [Google Scholar]
- 25.Gayathri K. S., Elias S., Ravindran B. Hierarchical activity recognition for dementia care using Markov Logic Network. Personal and Ubiquitous Computing. 2015;19(2):271–285. doi: 10.1007/s00779-014-0827-7. [DOI] [Google Scholar]
- 26.Phua C., Roy P. C., Aloulou H., et al. State-of-the-art assistive technology for people with dementia. In: Chong N.-Y., Mastrogiovanni F., editors. Handbook of Research on Ambient Intelligence and Smart Environments: Trends and Perspectives. chapter 16. IGI Global; 2011. pp. 300–319. [DOI] [Google Scholar]
- 27.Roy P. C., Giroux S., Bouchard B., et al. Activity Recognition in Pervasive Intelligent Environments. Vol. 4. Berlin, Germany: Springer; 2011. A possibilistic approach for activity recognition in smart homes for cognitive assistance to Alzheimer's patients; pp. 33–58. (Atlantis Ambient and Pervasive Intelligence). [DOI] [Google Scholar]
- 28.Hildeman A. Classification of epileptic seizures using accelerometers [Ph.D. dissertation] Gothenburg, Sweden: Chalmers University of Technology; 2011. [Google Scholar]
- 29.Puyau M. R., Adolph A. L., Vohra F. A., Zakeri I., Butte N. F. Prediction of activity energy expenditure using accelerometers in children. Medicine and Science in Sports and Exercise. 2004;36(9):1625–1631. doi: 10.1249/01.mss.0000139898.30804.60. [DOI] [PubMed] [Google Scholar]
- 30.Gupta P., Dallas T. Feature selection and activity recognition system using a single triaxial accelerometer. IEEE Transactions on Biomedical Engineering. 2014;61(6):1780–1786. doi: 10.1109/TBME.2014.2307069. [DOI] [PubMed] [Google Scholar]
- 31.Özdemir A. T., Barshan B. Detecting falls with wearable sensors using machine learning techniques. Sensors. 2014;14(6):10691–10708. doi: 10.3390/s140610691. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Arif M., Kattan A. Physical activities monitoring using wearable acceleration sensors attached to the body. PLoS ONE. 2015;10(7) doi: 10.1371/journal.pone.0130851.e0130851 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Alvarez G. G., Ayas N. T. The impact of daily sleep duration on health: a review of the literature. Progress in Cardiovascular Nursing. 2004;19(2):56–59. doi: 10.1111/j.0889-7204.2004.02422.x. [DOI] [PubMed] [Google Scholar]
- 34.Crouter S. E., Churilla J. R., Bassett D. R., Jr. Estimating energy expenditure using accelerometers. European Journal of Applied Physiology. 2006;98(6):601–612. doi: 10.1007/s00421-006-0307-5. [DOI] [PubMed] [Google Scholar]
- 35.Yu H., Spenko M., Dubowsky S. An adaptive shared control system for an intelligent mobility aid for the elderly. Autonomous Robots. 2003;15(1):53–66. doi: 10.1023/A:1024488717009. [DOI] [Google Scholar]
- 36.Achlioptas D. Database-friendly random projections. Proceedings of the 20th ACM SIGMOD-SIGACT-SIGART Symposium on Principles of Database Systems (PODS '01); May 2001; Santa Barbara, Calif, USA. pp. 274–281. [Google Scholar]
- 37.Mathie M. J., Celler B. G., Lovell N. H., Coster A. C. F. Classification of basic daily movements using a triaxial accelerometer. Medical and Biological Engineering and Computing. 2004;42(5):679–687. doi: 10.1007/BF02347551. [DOI] [PubMed] [Google Scholar]
- 38.Zhang M., Sawchuk A. A. USC-HAD: a daily activity dataset for ubiquitous activity recognition using wearable sensors. Proceedings of the International Conference on Ubiquitous Computing (UbiComp '12); September 2012; ACM; pp. 1036–1043. [DOI] [Google Scholar]
- 39.Atallah L., Lo B., King R., Yang G.-Z. Sensor positioning for activity recognition using wearable accelerometers. IEEE Transactions on Biomedical Circuits and Systems. 2011;5(4):320–329. doi: 10.1109/TBCAS.2011.2160540. [DOI] [PubMed] [Google Scholar]
- 40.Bayat A., Pomplun M., Tran D. A. A study on human activity recognition using accelerometer data from smartphones. Procedia Computer Science. 2014;34:450–457. doi: 10.1016/j.procs.2014.07.009. [DOI] [Google Scholar]
- 41.Berchtold M., Budde M., Gordon D., Schmidtke H. R., Beigl M. ActiServ: activity recognition service for mobile phones. Proceedings of the 14th IEEE International Symposium on Wearable Computers (ISWC '10); October 2010; Seoul, South Korea. IEEE; pp. 1–8. [DOI] [Google Scholar]
- 42.Henpraserttae A., Thiemjarus S., Marukatat S. Accurate activity recognition using a mobile phone regardless of device orientation and location. Proceedings of the International Conference on Body Sensor Networks (BSN '11); May 2011; Dallas, Tex, USA. IEEE; pp. 41–46. [DOI] [Google Scholar]
- 43.Hoque E., Stankovic J. AALO: activity recognition in smart homes using Active Learning in the presence of Overlapped activities. Proceedings of the 6th International Conference on Pervasive Computing Technologies for Healthcare and Workshops (PervasiveHealth '12); May 2012; pp. 139–146. [DOI] [Google Scholar]
- 44.Iso T., Yamazaki K. Gait analyzer based on a cell phone with a single three-axis accelerometer. Proceedings of the 8th International Conference on Human-Computer Interaction with Mobile Devices and Services (MobileHCI '06); September 2006; Espoo, Finland. pp. 141–144. [DOI] [Google Scholar]
- 45.Kose M., Incel O. D., Ersoy C. Online human activity recognition on smart phones. Proceedings of the Workshop on Mobile Sensing: From Smartphones and Wearables to Big Data (Colocated with IPSN); April 2012; Beijing, China. pp. 11–15. [Google Scholar]
- 46.Kwapisz J. R., Weiss G. M., Moore S. A. Activity recognition using cell phone accelerometers. ACM SIGKDD Explorations Newsletter. 2011;12(2):74–82. doi: 10.1145/1964897.1964918. [DOI] [Google Scholar]
- 47.Lane N., Mohammod M., Lin M., et al. Bewell: a smartphone application to monitor, model and promote wellbeing. Proceedings of the 5th International ICST Conference on Pervasive Computing Technologies for Healthcare; 2012; IEEE; pp. 23–26. [Google Scholar]
- 48.Lee Y. S., Cho S. Hybrid Artificial Intelligent Systems: 6th International Conference, HAIS 2011, Wroclaw, Poland, May 23–25, 2011, Proceedings, Part I. Vol. 6678. Berlin, Germany: Springer; 2011. Activity recognition using hierarchical hidden markov models on a smartphone with 3d accelerometer; pp. 460–467. (Lecture Notes in Computer Science). [DOI] [Google Scholar]
- 49.Mannini A., Sabatini A. M. Machine learning methods for classifying human physical activity from on-body accelerometers. Sensors. 2010;10(2):1154–1175. doi: 10.3390/s100201154. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Maurer U., Smailagic A., Siewiorek D. P., Deisher M. Activity recognition and monitoring using multiple sensors on different body positions. Proceedings of the International Workshop on Wearable and Implantable Body Sensor Networks (BSN '06); April 2006; Cambridge, Mass, USA. IEEE; pp. 113–116. [DOI] [Google Scholar]
- 51.Miluzzo E., Lane N. D., Fodor K., et al. Sensing meets mobile social networks: the design, implementation and evaluation of the CenceMe application. Proceedings of the 6th ACM Conference on Embedded Networked Sensor Systems (SenSys '08); November 2008; Raleigh, NC, USA. pp. 337–350. [DOI] [Google Scholar]
- 52.Pärkkä J., Ermes M., Korpipää P., Mäntyjärvi J., Peltola J., Korhonen I. Activity classification using realistic data from wearable sensors. IEEE Transactions on Information Technology in Biomedicine. 2006;10(1):119–128. doi: 10.1109/TITB.2005.856863. [DOI] [PubMed] [Google Scholar]
- 53.Saponas T. S., Lester J., Froehlich J., Fogarty J., Landay J. iLearn on the iPhone: Real-Time Human Activity Classification on Commodity Mobile Phones. 2008. [Google Scholar]
- 54.Siirtola P., Röning J. Recognizing human activities user-independently on smartphones based on accelerometer data. International Journal of Interactive Multimedia and Artificial Intelligence. 2012;1(5):38–45. doi: 10.9781/ijimai.2012.155. [DOI] [Google Scholar]
- 55.Sohn T., Varshavsky A., LaMarca A., et al. UbiComp 2006: Ubiquitous Computing: 8th International Conference, UbiComp 2006 Orange County, CA, USA, September 17–21, 2006 Proceedings. Vol. 4206. Berlin, Germany: Springer; 2006. Mobility detection using everyday GSM traces; pp. 212–224. (Lecture Notes in Computer Science). [DOI] [Google Scholar]
- 56.Yang J. Toward physical activity diary: motion recognition using simple acceleration features with mobile phones. Proceedings of the 1st International Workshop on Interactive Multimedia for Consumer Electronics (IMCE '09) 2009 doi: 10.1145/1631040.1631042. [DOI] [Google Scholar]
- 57.Zhu C., Sheng W. Motion- and location-based online human daily activity recognition. Pervasive and Mobile Computing. 2011;7(2):256–269. doi: 10.1016/j.pmcj.2010.11.004. [DOI] [Google Scholar]
- 58.Liu H., Yu L. Toward integrating feature selection algorithms for classification and clustering. IEEE Transactions on Knowledge and Data Engineering. 2005;17(4):491–502. doi: 10.1109/TKDE.2005.66. [DOI] [Google Scholar]
- 59.Robnik-Šikonja M., Kononenko I. Theoretical and empirical analysis of ReliefF and RReliefF. Machine Learning. 2003;53(1-2):23–69. doi: 10.1023/a:1025667309714. [DOI] [Google Scholar]
- 60.Hall M. A. Correlation-based feature selection for machine learning [Ph.D. thesis] The University of Waikato; 1999. [Google Scholar]
- 61.Jain A., Zongker D. Feature selection: evaluation, application, and small sample performance. IEEE Transactions on Pattern Analysis and Machine Intelligence. 1997;19(2):153–158. doi: 10.1109/34.574797. [DOI] [Google Scholar]
- 62.Guyon I., Elisseeff A. An introduction to variable and feature selection. Journal of Machine Learning Research. 2003;3:1157–1182. [Google Scholar]
- 63.Pirttikangas S., Fujinami K., Nakajima T. Feature selection and activity recognition from wearable sensors. In: Youn H. Y., Kim M., Morikawa H., editors. Ubiquitous Computing Systems. Vol. 4239. Berlin, Germany: Springer; 2006. pp. 516–527. (Lecture Notes in Computer Science). [DOI] [Google Scholar]
- 64.Johnson W. B., Lindenstrauss J. Conference in Modern Analysis and Probability. Vol. 26. American Mathematical Society; 1984. Extensions of Lipshitz mapping into Hilbert space; pp. 189–206. (Contemporary Mathematics). [Google Scholar]
- 65.Parzen E. On estimation of a probability density function and mode. The Annals of Mathematical Statistics. 1962;33(3):1065–1076. doi: 10.1214/aoms/1177704472. [DOI] [Google Scholar]
- 66.Zheng Y. Human activity recognition based on the hierarchical feature selection and classification framework. Journal of Electrical and Computer Engineering. 2015;2015:9. doi: 10.1155/2015/140820.140820 [DOI] [Google Scholar]
- 67.Sivakumar A. Geometry aware compressive analysis of human activities: application in a smart phone platform [M.S. thesis] Tempe, Ariz, USA: Arizona State University; 2014. [Google Scholar]
- 68.Vaka P. R. A pervasive middleware for activity recognition with smartphones [M.S. thesis] University of Missouri-Kansas; 2015. [Google Scholar]
- 69.Tran D., Sorokin A. Human activity recognition with metric learning. In: Forsyth D., Torr P., Zisserman A., editors. Computer Vision—ECCV 2008. Vol. 5302. Berlin, Germany: Springer; 2008. pp. 548–561. (Lecture Notes in Computer Science). [DOI] [Google Scholar]
- 70.Lee D., Yang M.-H., Oh S. Fast and accurate head pose estimation via random projection forests. Proceedings of the IEEE International Conference on Computer Vision (ICCV '15); December 2015; Santiago, Chile. IEEE; pp. 1958–1966. [DOI] [Google Scholar]
- 71.Zhao L., Wang X., Sukthankar G., Sukthankar R. Motif discovery and feature selection for CRF-based activity recognition. Proceedings of the 20th International Conference on Pattern Recognition (ICPR '10); August 2010; Istanbul, Turkey. pp. 3826–3829. [DOI] [Google Scholar]