Abstract
Many intelligent systems that focus on the needs of a human require information about the activities being performed by the human. At the core of this capability is activity recognition, which is a challenging and well-researched problem. Activity recognition algorithms require substantial amounts of labeled training data yet need to perform well under very diverse circumstances. As a result, researchers have been designing methods to identify and utilize subtle connections between activity recognition datasets, or to perform transfer-based activity recognition. In this paper we survey the literature to highlight recent advances in transfer learning for activity recognition. We characterize existing approaches to transfer-based activity recognition by sensor modality, by differences between source and target environments, by data availability, and by type of information that is transferred. Finally, we present some grand challenges for the community to consider as this field is further developed.
Keywords: Machine Learning, Activity Recognition, Transfer Learning, Smart Environments
1. Introduction
Researchers in the artificial intelligence community have struggled for decades trying to build machines capable of matching or exceeding the mental capabilities of humans. One capability that continues to challenge researchers is designing systems which can leverage experience from previous tasks into improved performance in a new task which has not been encountered before. When the new task is drawn from a different population than the old, this is considered to be transfer learning. The benefits of transfer learning are numerous; less time is spent learning new tasks, less information is required of experts (usually human), and more situations can be handled effectively. These potential benefits have lead researchers to apply transfer learning techniques to many domains with varying degrees of success.
One particularly interesting domain for transfer learning is human activity recognition. The goal of human activity recognition is to be able to correctly classify the current activity a human or group of humans is performing given some set of data. Activity recognition is important to a variety of applications including health monitoring, automatic security surveillance, and home automation. As research in this area has progressed, an increasing number of researchers have started looking at ways transfer learning can be applied to reduce the training time and effort required to initialize new activity recognition systems, to make the activity recognition systems more robust and versatile, and to effectively reuse the existing knowledge that has previously been generated.
With the recent explosion in the number of researchers and the amount of research being done on transfer learning, activity recognition, and transfer learning for activity recognition, it becomes increasingly important to critically analyze this body of work and discover areas which still require further investigation. Although recent progress in transfer learning has been analyzed in (Pan and Yang, 2010; Taylor and Stone, 2009; Vilalta and Drissi, 2002) and several surveys have been conducted on activity recognition (Alemdar and Ersoy, 2010; Avci, Bosch, Marin-Perianu, Marin-Perianu and Havinga, 2010; Chan, Estve, Escriba and Campo, 2008; Haigh and Yanco, 2002) no one has specifically looked into the intersection of these two areas. This survey, therefore, examines the field of transfer-based activity recognition and the unique challenges presented in this domain. For an overview of the survey, see Figure 1 which illustrates the topics covered in this survey and how they relate to each other.
2. Background
Activity recognition aims to identify activities as they occur based on data collected by sensors. There exist a number of approaches to activity recognition (Kim, Helal and Cook, 2010) that vary depending on the underlying sensor technologies that are used to monitor activities, the alternative machine learning algorithms that are used to model the activities and the realism of the testing environment.
Advances in pervasive computing and sensor networks have resulted in the development of a wide variety of sensor modalities that are useful for gathering information about human activities. Wearable sensors such as accelerometers are commonly used for recognizing ambulatory movements (e.g., walking, running, sitting, climbing, and falling) (Krishnan and Panchanathan, 2008; Maurer, Smailagic, Siewiorek and Deisher, 2006). More recently, researchers are exploring smart phones equipped with accelerometers and gyroscopes to recognize such movement and gesture patterns (Kwapisz, Weiss and Moore, 2010).
Environment sensors such as infrared motion detectors or magnetic door sensors have been used to gather information about more complex activities such as cooking, sleeping, and eating. These sensors are adept in performing location-based activity recognition in indoor environments (Agrawal and Srikant, 1995; Logan, Healey, Philipose, Tapia and Intille, 2007; Kasteren, Englebienne and Kröse, 2010) just as GPS is used for outdoor environments (Liao, Fox and Kautz, 2005). Some activities such as washing dishes, taking medicine, and using the phone are characterized by interacting with unique objects. In response, researchers have explored the usage of RFID tags and shimmer sensors for tagging these objects and using the data for activity recognition (Palmes, Pung, Gu, Xue and Chen, 2010; Philipose, P, Perkowitz, Patterson, Fox, Kautz and Hhnel, 2004). Researchers have also used data from video cameras and microphones as well (Agrawal and Srikant, 1995).
There have been many varied machine learning models that have been used for activity recognition. These can be broadly categorized into template matching/transductive techniques, generative, and discriminative approaches. Template matching techniques employ a kNN classifier based on Euclidean distance or dynamic time warping. Generative approaches such as näive Bayes classifiers where activity samples are modeled using Gaussian mixtures have yielded promising results for batch learning. Generative probabilistic graphical models such as hidden Markov models and dynamic Bayesian networks have been used to model activity sequences and to smooth recognition results of an ensemble classifier (Lester, Choudhury, Kern, Borriello and Hannaford, 2005). Decision trees as well as bagging and boosting methods have been tested (Maurer et al., 2006). Discriminative approaches, including support vector machines and conditional random fields, have also been effective (Chang and Lin, 2011; Kasteren et al., 2010) and unsupervised discovery and recognition methods have also been introduced (Rashidi, Cook, Holder and Schmitter-Edgecombe, 2011; Gu, Chen, Tao and Lu, 2010). The traditional approaches to activity recognition make the strong assumption that the training and test data are drawn from identical distributions. Many real-world applications cannot be represented in this setting and thus the baseline activity recognition approaches have to be modified to work in these realistic settings. Transfer based activity recognition is one conduit for achieving this.
2.1. Transfer Learning
The ability to identify deep, subtle connections, what we term transfer learning, is the hallmark of human intelligence. Byrnes (Byrnes, 1996) defines transfer learning as the ability to extend what has been learned in one context to new contexts. Thorndike and Woodworth (Thorndike and Woodworth, 1901) first coined this term as they explored how individuals transfer learned concepts between contexts that share common features. Barnett and Ceci provide a taxonomy of features that influence transfer learning in humans (Barnett and Ceci, 2002).
In the field of machine learning, transfer learning is studied under a variety of different names including learning to learn, life-long learning, knowledge transfer, inductive transfer, context-sensitive learning, and metalearning (Arnold, Nallapati and Cohen, 2007; Elkan, 2001; Thrun, 1996; Thrun and Pratt, 1998; Vilalta and Drissi, 2002). It is also closely related to several other areas of machine learning such as self-taught learning, multi-task learning, domain adaptation, and co-variate shift. Because of this broad variance in the terms used to describe transfer learning it is helpful to provide a formal definition of transfer learning terms and of transfer learning itself which will be used throughout the rest of this paper.
2.2. Definitions
This survey starts with a review of basic definitions needed for discussions of transfer learning as it can be applied to activity recognition. Definitions for domain and task have been provided by Pan and Yang (Pan and Yang, 2010):
Definition 2.1 (Domain)
A domain D is a two-tuple (χ, P(X)). χ is the feature space of D and P(X) is the marginal distribution where X = {x1, …, xn} ∈ χ.
Definition 2.2 (Task)
A task T is a two-tuple (Y, f()) for some given domain D. Y is the label space of D and f() is an objective predictive function for D. f() is sometimes written as a conditional probability distribution P(y|x). f() is not given, but can be learned from the training data.
To illustrate these definitions, consider the problem of activity recognition using motion sensors. The domain is defined by a feature space which may represent the n-dimensional space defined by n sensor firing counts within a given time window and a marginal probability distribution over all possible firing counts. The task is composed of a label space y which consists of the set of labels for activities of interest, and a conditional probability distribution consisting of the probability of assigning a label yi ∈ y given the observed instance x ∈ χ.
Using these terms, we can now define transfer learning. In this paper we specify a definition of transfer learning that is similar to that presented by Pan and Yang (Pan and Yang, 2010) but we allow for transfer learning which uses multiple source domains.
Definition 2.3 (Transfer Learning)
Given a set of source domains DS = Ds1, …, Dsn where n > 0, a target domain, Dt, a set of source tasks TS = Ts1, …Tsn where Tsi ∈ TS corresponds with Dsi ∈ DS, and a target task Tt which corresponds to Dt, transfer learning helps improve the learning of the target predictive function ft() in Dt where Dt ∉ DS and Tt ∉ TS.
This definition of transfer learning is broad and encompasses a large number of different transfer learning scenarios. The source domains can differ from the target domain by having a different feature space, a different distribution of instances in the feature space, or both. The source tasks can differ from the target task by having a different label space, a different predictive function for labels in that label space, or both. The source data can differ from the target data by having a different domain, a different task, or both. However, all transfer learning problems rely on the basic assumption that there exists some relationship between the source and target areas which allows for the successful transfer of knowledge from the source to the target.
2.3. Scenarios
To further illustrate the variety of problems which fall under the scope of transfer-based activity recognition, we provide illustrative scenarios. Not all of these scenarios can be addressed by current transfer learning methods. The first scenario represents a typical transfer learning problem solvable using recently developed techniques. The second scenario represents a more challenging situation that pushes the boundaries of current transfer learning techniques. The third scenario requires a transfer of knowledge across such a large difference between source and target datasets, that current techniques only scratch the surface of what is required to make such a knowledge transfer successful.
2.3.1. Scenario 1
In one home which has been equipped with multiple motion and temperature sensors, an activity recognition algorithm has been trained using months of annotated labels to provide the ground truth for activities which occur in that home. A transfer learning algorithm should be able to reuse the labeled data to perform activity recognition in a new setting. Such transfer will save months of man-hours annotating data for the new home. However, the new home has a different layout as well as a different resident and different sensor locations than the first home.
2.3.2. Scenario 2
An individual with Parkinson’s disease visits his neurosurgeon twice a year to get an updated assessment of his gait, tremor, and cognitive health. The medical sta3 perform some gait measurements and simulated activities in their office space to determine the effectiveness of the prescribed medication, but want to determine if the observed improvement is reflected in the activities the patient performs in his own home. A learning algorithm will need to be able to transfer information between different physical settings, as well as time of day, sensors used, and scope of the activities.
2.3.3. Scenario 3
A researcher is interested in studying the cooking activity patterns of college students living in university dorms in the United States. The research study has to be conducted using the smart phone of the student as the sensing mechanism. The cooking activity of these students typically consists of heating up a frozen snack from the refrigerator in the microwave oven. In order to build the machine learning models for recognizing these activity patterns, the researcher has access to cooking activities for a group of grandmothers living in India. This dataset was collected using smart home environmental sensors embedded in the kitchen and the cooking activity itself was very elaborate. Thus the learning algorithm is now faced with changes in the data at many layers; namely, differences in the sensing mechanisms, cultural changes, age related differences, different location settings and finally differences in the activity labels. This transfer learning from one setting to another diverse setting is most challenging and requires significant progress in transfer learning domain to even attempt to solve the problem.
These scenarios illustrate different types of transfer that should be possible using machine learning methods for activity recognition. As is described by these situations, transfer may occur across several dimensions. We next take a closer look at these types of transfer and use these descriptors to characterize existing approaches to transfer learning for activity recognition.
2.4. Dimensions of Analysis
Transfer learning can take many forms in the context of activity recognition. In this discussion we consider four dimensions to characterize various approaches to transfer learning for activity recognition. First, we consider different sensor modalities on which transfer learning has been applied. Second, we consider differences between the source and target environments in which data is captured. The third dimension is the amount and type of data labeling that is available in source and target domains. Finally, we examine the representation of the knowledge that is transferred from source to target. The next sections discuss these dimensions in more detail and characterize existing work based on alternative approaches to handling such differences.
3. Modality
One natural method for the classification of transfer learning techniques is the underlying sensing modalities used for activity recognition. Some techniques may be generalizable to different sensor modalities, but most techniques are too specific to be generally applicable to any sensor modality other than that for which they are designed to work with. This is usually because the types of differences that occur between source and target domains are different for each sensor modality. These differences and their effect on the transfer learning technique are discussed in detail in Section 4. In this section we consider only those techniques which have empirically demonstrated their ability to operate on a given sensor modality.
The classification of sensor modalities itself is a difficult problem and indeed creating precise classification topology is outside of the scope of this paper. However, we roughly categorize sensor modalities into the following classifications, video cameras, wearable devices, and ambient sensors. For each sensor modality, we provide a brief description of the types of sensors which are included and a summary of the research works performing transfer learning in that domain. In this section we do not describe the transfer learning algorithms used in the papers as that will be discussed in the other dimensions of analysis.
3.1. Video Sequences
Video cameras are one of the first sensor modalities in which transfer learning has been applied to the area of activity recognition (Yang, Yan and Hauptmann, 2007). Video cameras provide a dense feature-space for activity recognition which potentially allows for extremely fine-grained recognition of activities. Spatio-temporal features are extracted from video sequences for characterizing the activities occurring in them. Activity models are then learned using these feature representations.
One drawback of video processing for activity recognition is that the use of video cameras raises more issues associated with user privacy. In addition, cameras need to be well positioned and track individuals in order to capture salient data for processing. Activity recognition via video cameras has received broad attention in transfer learning research (Duan, Xu, Tsang and Luo, 2010; Farhadi and Tabrizi, 2008; Lam, Roy-Chowdhury and Shelton, 2011; Liu, Shah, Kuipers and Savarese, 2011; Nater, Tommasi, Grabner, van Gool and Caputo, 2011; Wei and Pal, 2011; Wu, Khalili and Aghajan, 2010; Xianming and Shaozi, 2009; Yang et al., 2007; Yang, Wang and Mori, 2011).
3.2. Wearable sensors
Body Sensor Networks are another commonly used sensing mechanism to capture activity related information from individuals. These sensors are typically worn by the individuals. Strategic placement of the sensors helps in capturing important activity related information such as movements of the upper and lower parts of the body that can then be used to learn activity models. Sensors in this category include, inertial sensors such as accelerometers and gyroscopes, sensors embedded in smart phones, radio frequency identification sensors and tags. Researchers have applied transfer learning techniques to both activity recognition using wearable accelerometers and activity recognition using smartphones but we have not seen any transfer learning approaches applied to activity recognition using RFID tags. This may be due in part to the relatively low use of RFID tags in activity recognition itself.
Within wearable sensors, two types of problems are generally considered. The first is the problem of activity recognition itself (Blanke and Schiele, 2010; Calatroni, Roggen and Tröster, 2011; Chieu, Lee and Kaelbling, 2006; Hachiya, Sugiyama and Ueda, 2012; Krishnan, Lade and Panchanathan, 2010; Kurz, Hölzl, Ferscha, Calatroni, Roggen and Tröster, 2011; Roggen, Frster, Calatroni and Trster, 2011; Venkatesan, Krishnan and Panchanathan, 2010; Zhao, Chen, Liu and Liu, 2010; Zhao, Chen, Liu, Shen and Liu, 2011), and the second is the problem of user localization, which can then be used to increase the accuracy of the activity recognition algorithm (Pan, Tsang, Kwok and Yang, 2011; Pan, Kwok, Yang and Pan, 2007; Pan, Shen, Yang and Kwok, 2008; Pan, Zheng, Yang and Hu, 2008; Zheng, Pan, Yang and Pan, 2008). Both problems present interesting challenges for transfer learning.
3.3. Ambient Sensors
Ambient sensors represent the broadest classification of sensor modalities which we define in this paper. We categorize any sensor that is neither wearable nor video camera into ambient sensors. These sensors are typically embedded in an individual’s environment. This category includes a wide variety of sensors such as motion detectors, door sensors, object sensors, pressure sensors, and temperature sensors. As the name indicates, these sensors collect a variety of activity related information such as human movements in the environment induced by activities, interactions with objects during the performance of an activity, and changes to illumination, pressure and temperature in the environment due to activities. Researchers have only recently begun to look at transfer learning applications for ambient sensors with the earliest work appearing around 2008 (van Kasteren, Englebienne and Kröse, 2008). Since then the field of transfer learning for activity recognition using ambient sensors has progressed rapidly with many different research groups analyzing the problem from several different angles (Cook, 2010; Hu and Yang, 2011; Rashidi and Cook, 2009; Rashidi and Cook, 2010a; Rashidi and Cook, 2010b; Rashidi and Cook, 2011; Roggen et al., 2011; van Kasteren, Englebienne and Krse, 2010; Zheng, Hu and Yang, 2009).
3.4. Crossing the sensor boundaries
Clearly, transfer learning within individual sensor modalities is progressing. Researchers are actively developing and applying new techniques to solve a variety of problems within any given sensor modality domain. However, there has been little work done that tries to transfer knowledge between any two or more sensor modalities. Kurz et al. (Kurz et al., 2011) and Roggen et al. (Roggen et al., 2011) address this problem using a teacher/learner model which is discussed further in Section 5. Hu et al. (Hu and Yang, 2011) introduce a transfer learning technique for successfully transferring some knowledge across sensor modalities, but greater transfer of knowledge between modalities has yet to be explored.
4. Physical Setting Differences
Another useful categorization of transfer learning techniques is the types of physical differences between a source and target dataset across which the transfer learning techniques can achieve a successful transfer of knowledge. In this section, we describe these differences in a formal setting and provide illustrative examples drawn from activity recognition.
We use the terminology for domain, task and transfer learning defined in Section 2 to describe the differences between source and target datasets. These differences can be in the form of the feature-space representation, the marginal probability distribution of the instances, the label space, and/or the objective predictive function. When describing transfer learning in general, using such broad terms allows one to encompass many different problems. However, when describing transfer learning for a specific application, such as activity recognition, it is convenient to use more application specific terms. For example, differences in the feature-space representation can be thought of in terms of the sensor modalities and sampling rates and differences in the marginal probability distribution can be thought of in terms of different people performing the same activity, or having the activity performed in different physical spaces.
Even when limiting the scope to activity recognition, it is still infeasible to enumerate every possible difference between source and target datasets. In this survey we consider some of the most common or important differences between the source and target datasets including time, people, devices, space, sensor types, and labels. Table 1 summarizes the relationship between each of these applied differences and the formal definitions of transfer learning differences.
Table 1.
Formal Definition | Applied Meaning |
---|---|
χt ≠ χsi for 0 < i < n | sensor networks, sensor modality, or physical space |
P(Xt) ≠ P(Xsi) for 0 < i < n | time, people, devices, or sampling rates |
Yt ≠ Ysi for 0 < i < n | activities or labels |
ft(x) ≠ fsi (x) for 0 < i < n | time, people, devices, sampling rates, activities, or labels |
Differences across time, people, devices, or sensor sampling rates result in differences in the underlying marginal probability distribution, the objective predictive function, or both. Several papers focus specifically on transferring across time differences (Krishnan et al., 2010; Pan et al., 2011; Pan et al., 2007; Venkatesan et al., 2010), differences between people (Chattopadhyay, Krishnan and Panchanathan, 2011; Hachiya et al., 2012; Rashidi and Cook, 2009; Zhao et al., 2011), and differences between devices (Zhao et al., 2010; Zheng et al., 2008).
Differences created when comparing datasets from different spaces or spatial layouts are reflected by differences in the feature-spaces, the marginal probability distributions, the objective predictive functions, or any combination of these. As the number of differences increases, the source and target datasets become less related making transfer learning more difficult. Because of this, current research usually imposes limiting assumptions about what is different between the spaces. Several researchers, for example, assume that some meta-features are added which provide space-independent information (Cook, 2010; Rashidi and Cook, 2010a; Rashidi and Cook, 2010b; Rashidi and Cook, 2011; van Kasteren et al., 2008; van Kasteren et al., 2010). For WiFi localization, Pan et al. (Pan, Shen, Yang and Kwok, 2008) assume that the source and target spaces are in the same building. Applying transfer learning to video clips from different spaces usually results in handling issues of background differences (Cao, Liu and Huang, 2010; Xianming and Shaozi, 2009; Yang et al., 2007) and/or issues of camera view angle (Liu et al., 2011).
Differences in the labels used in the datasets are obviously reflected by differences in the label space and the objective predictive function. Compared to the other differences discussed previously, transferring between differences in the label space has received much less attention in the current literature (Hu and Yang, 2011; Lam et al., 2011; Wei and Pal, 2011; Yang et al., 2011; Zheng et al., 2009).
One of the largest differences between datasets occurs when the source and target datasets have a different sensor modality. This makes the transfer learning problem much more difficult and relatively little work has been done in this direction. Hu and Yang have started work in this direction in (Hu and Yang, 2011). Additionally, Calatroni et al. (Calatroni et al., 2011), Kurz et al. (Kurz et al., 2011) and Roggen et al. (Roggen et al., 2011) take a different approach to transferring across sensor modality by assuming a classifier for the source modalities can act as an expert for training a classifier in the target sensor modality.
5. Data Labeling
In this section we consider the problem of transfer learning from the perspective of the availability of labeled data. Traditional machine learning uses the terms supervised learning and unsupervised learning to distinguish learning techniques based on the availability and use of labeled data. To distinguish between source and target labeled data availability we introduce two new terms, informed and uninformed, which we apply to the availability of labeled data in the target area. Thus, informed supervised (IS) transfer learning implies that some labeled data is available in both the target and source domains. Uninformed supervised (US) transfer learning implies that labeled data is available only in the source domain. Informed unsupervised (IU) transfer learning implies that labeled data is only available in the target domain. Finally, uninformed unsupervised (UU) transfer learning implies that no labeled data is available for either the source or target domains. One final case to consider is teacher/learner (TL) transfer learning, where no training data is directly available. Instead a previously-trained classifier (the Teacher) is introduced which operates simultaneously with the new classifier to be trained (the Learner) and provides the labels for observed data instances.
Two other terms that are often used in machine learning literature and may be applicable here are inductive and transductive learning. Inductive learning refers to learning techniques which try to learn the objective predictive function. Transductive learning techniques, on the other hand, try to learn the relationship between instances. Pan and Yang (Pan and Yang, 2010) extend the definitions of inductive and transductive learning to transfer learning, but the definitions do not create a complete taxonomy for transfer learning techniques. For this reason, we do not specifically classify recent works as being inductive or transductive in nature, but we note here how the inductive and transductive definitions fit into a classification based upon the availability of labeled data.
Inductive learning requires that labeled data be available in the target domain regardless of its availability in the source domain. Thus, most informed supervised and informed unsupervised transfer learning techniques are also inductive transfer learning techniques. Transductive learning, however, does not require labeled data in the target domain. Therefore, most uninformed supervised techniques are also transductive transfer learning techniques. Table 2 summarizes this general relationship.
Table 2.
Label Availability | Most Common Approach |
---|---|
Informed Supervised | Inductive Learning |
Informed Unsupervised | Inductive Learning |
Uninformed Supervised | Transductive Learning |
Uninformed Unsupervised | Unsupervised Learning |
Several researchers have developed and applied informed, supervised transfer learning techniques for activity recognition. These techniques have been applied to activity recognition using wearables (Blanke and Schiele, 2010; Krishnan, 2010; Pan, Yang, Chang and Yeung, 2006; Pan et al., 2007; Venkatesan et al., 2010; Zheng et al., 2008) and to activity recognition using cameras (Duan et al., 2010; Lam et al., 2011; Nater et al., 2011; Xianming and Shaozi, 2009; Yang et al., 2007; Yang et al., 2011).
Research into transfer-based activity recognition using ambient sensors has almost exclusively focused on uninformed supervised transfer learning (Cook, 2010; Hu and Yang, 2011; Hu, Zheng and Yang, 2010; Rashidi and Cook, 2009; van Kasteren et al., 2008; van Kasteren et al., 2010; Venkatesan et al., 2010; Zheng et al., 2009), but a few algorithms are able to take advantage of the labeled target data if it is available (Rashidi and Cook, 2010a; Rashidi and Cook, 2010b; Rashidi and Cook, 2011). This focus on uninformed supervised transfer learning is most likely due to the allurement of building an activity recognition framework that can be trained offline and later installed into any user’s space without requiring additional data labeling effort. Wearables have also been used for uninformed supervised transfer learning research (Hachiya et al., 2012; Pan et al., 2011; Pan, Shen, Yang and Kwok, 2008; Pan, Zheng, Yang and Hu, 2008; Yang, 2009; Zhao et al., 2010; Zhao et al., 2011) as have cameras(Cao et al., 2010; Farhadi and Tabrizi, 2008; Liu et al., 2011; Wei and Pal, 2011; Wu et al., 2010).
Despite the abundance of research using labeled source data, research into transfer learning techniques for activity recognition in which no source labels are available is extremely sparse. Pan et al. (Pan et al., 2011) have applied an uninformed unsupervised technique, transfer component analysis (TCA) to reduce the distance between domains by learning some transfer components across domains in a reproducing kernel Hilbert space using maximum mean discrepancy. We are unaware of any other work for uninformed unsupervised transfer-based activity recognition. We are also unaware of any work on informed unsupervised transfer-based activity recognition. The lack of research into informed unsupervised transfer-based activity recognition is not surprising because the idea of having labeled target data available and not having labeled source data is counterintuitive to the general principle of transfer learning. However, informed unsupervised transfer learning may still provide significant benefits to activity recognition.
The teacher/learner model for activity recognition is considerably less studied than the previously discussed techniques. However, we feel that this area has significant promise for improving transfer learning for activity recognition and making activity recognition systems much more robust and versatile. Roggen et al. (Roggen et al., 2011), Kurz et al. (Kurz et al., 2011), and Calatroni et al. (Calatroni et al., 2011) apply the teacher/learner model to develop an opportunistic system which is capable of using whatever sensors are currently contained in the environment to perform activity recognition.
In order for the teacher/learner model to be applicable, two requirements must be met. First, an existing classifier (the teacher) must already be trained in the source domain. Second, the teacher must operate simultaneously with a new classifier in the target domain (the learner) to provide the training for the learner. For example, Roggen et al. (Roggen et al., 2011) equip a cabinet of drawers with an accelerometer for each drawer and then a classifier is trained to recognize which drawer of the cabinet is being opened or closed. This classifier becomes the teacher. Then several wearable accelerometers are attached to the person opening and closing the drawers. Now, a new classifier is trained using the wearable accelerometers. This classifier is the learner. When the individual opens or closes a drawer, the teacher labels the activity according to its classification model. This label is given to the learner which can then be used as labeled training data in real-time without the need to supply any manually labeled data.
The teacher/learner model presents a new perspective on transfer learning and introduces additional challenges. One major challenge of the teacher/learner model is that the accuracy of the learner is limited by the accuracy of the teacher. Additionally, the system’s only source of a ground truth comes from the teacher and thus the learner is completely reliant upon the teacher. It remains to be explored whether the learner can ever outperform the teacher and if it does so, whether it can convince itself and others of this superior performance. Finally, while the teacher/learner model provides a convenient way to transfer across different domains, an additional transfer mechanism would need to be employed to transfer across different label spaces.
6. Type of Knowledge Transferred
Pan and Yang (Pan and Yang, 2010) describe four general classifications for transfer learning in relation to what is transferred, instance transfer, feature-representation transfer, parameter transfer, and relational-knowledge transfer.
6.1. Instance Transfer
Instance transfer reuses the source data to train the target classifier, usually by re-weighting the source instances based upon a given metric. Instance transfer techniques work well when χs = χt i.e., the feature space describing the source and target domains are same. They may also be applied after the feature representation has first been transferred to a common representation between the source and target domains.
Several researchers have applied instance transfer techniques to activity recognition. Hachiya et al. (Hachiya et al., 2012) develop an importance weighted least-squares probabilistic classification approach to handle transfer learning when P(Xs) ≠ P(Xt) (i.e., the co-variate shift problem) and apply this approach to wearable accelerometers. Venkatesan et al. (Krishnan, 2010; Venkatesan, 2011; Venkatesan et al., 2010) extend the AdaBoost framework proposed by Freund and Schapire (Freund and Schapire, 1997) to include cost-sensitive boosting which tries to weight samples from the source domain according to their relevance in the target domain. In their approach, samples from the source domain are first given a relevance cost. As the classifier is trained, those instances from the source domain with a high relevance must also be classified correctly. Xianming and Shaozi apply TrAdaBoost (a different transfer learning extension of AdaBoost) (Dai, Yang, Xue and Yu, 2007) to action recognition in video clips (Xianming and Shaozi, 2009). Lam et al. weight the source and target data differently when training an SVM to recognize target actions from video clips (Lam et al., 2011). Training a typical SVM involves solving the following optimization problem:
(1) |
where is the ith datapoint and yi, ξi are the label and slack variable associated with . w⃗ is the normal to the hyperplane. C is the parameter that trades o3 between training accuracy and margin size. However, to allow for the different source and target weights, they solve the following optimization:
(2) |
where the parameters are the same as before except the first n datapoints are from the source data and the last m datapoints are from the target data.
Unlike the previous instance-based approaches which weight the source instances based on similarity of features between the source and target data, Zheng et al. (Zheng et al., 2009) use an instance-based approach to weight source instances based upon the similarity between the label information of the source and target data. This allows them to transfer the labels from instances in the source domain to instances in the target domain using web-knowledge to relate the two domains (Hu and Yang, 2011; Hu et al., 2010). Taking a different approach, several researchers (Calatroni et al., 2011; Kurz et al., 2011; Roggen et al., 2011) use the real-time teacher/learner model discussed in the previous section to transfer the label of the current instance in the source domain to the instance in the target domain.
6.2. Feature-Representation Transfer
Feature-representation transfer reduces the differences between the source and target feature spaces. This can be accomplished by mapping the source feature space to the target feature space such as f : χs → χt, by mapping the target feature space to the source feature space such as g : χt → χs, or by mapping both the source and target feature spaces to a common feature space such as g : χt → χ and f : χs → χ. This mapping can be computed manually (van Kasteren et al., 2008) or learned as part of the transfer learning algorithm(Duan et al., 2010; Hu and Yang, 2011; Liu et al., 2011; Rashidi and Cook, 2009; Zheng et al., 2008).
When the mapping is part of the transfer learning algorithm a common approach is to apply a dimensionality reduction technique to map both source and target feature-space into a common latent space(Pan et al., 2006; Pan et al., 2011; Pan et al., 2007; Pan, Zheng, Yang and Hu, 2008). For example, Chattopadhyay et al. (Chattopadhyay et al., 2011) use Isomap (Tenenbaum, Silva and Langford, 2000) to map both the source and target data into a common low-dimensional space after which instance-based transfer techniques can be applied.
In some cases, meta-features are first manually introduced into the feature space and then the feature space is automatically mapped from the source domain to the target domain (Blanke and Schiele, 2010; Cook, 2010; van Kasteren et al., 2010). An example of this is the work of Rashidi and Cook (Rashidi and Cook, 2011). They first assign a location label to each sensor indicating in which room or functional area the sensor is located. Then activity templates are constructed from the data for both the source and target data, finally a mapping is learned between the source and target datasets based upon the similarity of activities and sensors (Rashidi and Cook, 2010a; Rashidi and Cook, 2010b).
6.3. Parameter Transfer
Parameter transfer learns parameters which are shared between the source and target tasks. One common use of parameter transfer is learning a prior distribution shared between the source and target datasets. For example, one technique (Cao et al., 2010) models the source and target tasks using a Gaussian Mixture Model which share a prior distribution, another algorithm (Duan et al., 2010) learns a target classifier using a set of pre-trained classifiers as prior for the target classifier, and van Kasteren et al. (van Kasteren et al., 2008) propose a method to learn the parameters of a Hidden Markov Model using labeled data from the source domain, and unlabeled data from the target domain. Later they extend this work to learn hyperparameter priors for the HMM instead of learning the parameters directly (van Kasteren et al., 2010).
Another common example of parameter transfer assumes the SVM parameter w can be split into two terms: w0, which is the same for both the source and target tasks, and v, which is specific to the particular task. Thus ws = w0 + vs and wt = w0 + vt. Several works adopt this approach (Nater et al., 2011; Yang et al., 2007).
Using a different approach to parameter transfer, a transfer learning algorithm (Pan, Shen, Yang and Kwok, 2008; Pan, Zheng, Yang and Hu, 2008) can extract knowledge from the source domain to impose additional constraints on a quadratically-constrained quadratic program optimization problem for the target domain. Along a similar line of thought, Zhao et al. (Zhao et al., 2010; Zhao et al., 2011) use information extracted from the source domain to initialize cluster centers for a k-means algorithm in the target domain.
6.4. Relational-Knowledge Transfer
Relational-knowledge transfer applies to problems in which the data is not independent and identically distributed (i.i.d.) as is traditionally assumed but can be represented through multiple relationships (Pan and Yang, 2010). Such problems are usually represented with a network or graph. Relational-knowledge transfer tries to transfer the relationships of in the source domain to the target domain. This type of transfer learning is not heavily explored, and as far as we are able to determine, no research is currently being pursued in transfer learning for activity recognition using relational-knowledge transfer.
7. Summary
The previous sections analyzed a large body of transfer-based activity recognition research along four different dimensions. Looking at each dimension separately provides an orderly way to analyze so many different papers. However, such separation may also make it difficult to see the bigger picture. Table 3, therefore, summarizes the classification of existing works along these four dimensions.
Table 3.
Paper | Sensor Modality | Difference | Labeling | Type of Knowledge Transfer |
---|---|---|---|---|
(Blanke and Schiele, 2010) | wearables | new activities and labels | IS | feature-representation |
(Calatroni et al., 2011) | wearables | different device, placement | TL | instance-based |
(Cao et al., 2010) | video camera | background, lighting, noise, and people | IS, US | parameter-based |
(Chattopadhyay et al., 2011) | wearables | people | IS | feature-representation and instance-based |
(Chieu et al., 2006) | wearables | people | IS | parameter-based |
(Cook, 2010) | ambient sensors | location, layout, people | US | feature-representation |
(Duan et al., 2010) | video camera | web-domain vs consumer domain. | IS | feature-representation and parameter-based |
(Farhadi and Tabrizi, 2008) | video camera | view angle | US | feature-space |
(Hachiya et al., 2012) | wearables | people | US | instance-based |
(Hu and Yang, 2011) | ambient sensors, wearables | label space, location | US | instance-based and feature-representation |
(Hu et al., 2010) | ambient sensors, wearables | label space | US | instance-based |
(Krishnan, 2010) | wearables | people and setting | IS | instance-based |
(Kurz et al., 2011) | wearables | sensors | TL | instance-based |
(Lam et al., 2011) | video camera | labels | IS | instance-based |
(Liu et al., 2011) | video camera | view angle | US | feature-representation |
(Nater et al., 2011) | video camera | activity sets, labels | IS | parameter-based |
(Pan et al., 2006) | wearables | time | IS | feature-representation |
(Pan et al., 2011) | wearables | time | US, UU | feature-representation |
(Pan et al., 2007) | wearables | time | IS | feature-representation |
(Pan, Shen, Yang and Kwok, 2008) | wearables | space, location | US | parameter-based |
(Pan, Zheng, Yang and Hu, 2008) | wearables | space, time, device | IS, US | feature-representation and parameter-based |
(Rashidi and Cook, 2009) | ambient sensors | people | US | feature-representation |
(Rashidi and Cook, 2010a) | ambient sensors | layout, sensor network | IS, US | feature-representation |
(Rashidi and Cook, 2010b) | ambient sensors | layout, sensor network | IS, US | feature-representation |
(Rashidi and Cook, 2011) | ambient sensors | layout, sensor network, people | IS, US | feature-representation |
(Roggen et al., 2011) | ambient sensors, wearables | devices | TL | instance-based |
(van Kasteren et al., 2008) | ambient sensors | location | US | feature-representation and parameter-based |
(van Kasteren et al., 2010) | ambient sensors | location | US | feature-representation and parameter-based |
(Venkatesan, 2011) | wearables | people, setting | IS | instance-based |
(Venkatesan et al., 2010) | wearables | people, setting | IS | instance-based |
(Wei and Pal, 2011) | video camera | labels | US | feature-representation |
(Wu et al., 2010) | video camera | view angle | US | parameter |
(Xian-ming and Shao-zi, 2009) | video camera | background, people | IS | instance |
(Yang et al., 2007) | video camera | background, video domain | IS | parameter-based |
(Yang, 2009) | wearables | space, time, device | IS, US | feature-representation and parameterbased |
(Yang et al., 2011) | video camera | activities performed | IS | distance function |
(Zhao et al., 2010) | wearables | mobile device, sampling rate | US | parameter-based |
(Zhao et al., 2011) | wearables | people | US | parameter-based |
(Zheng et al., 2009) | ambient sensors, wearables | activity labels | US | instance-based |
(Zheng et al., 2008) | wearables | devices | IS | feature-representation |
8. Grand Challenges
Although transfer-based activity recognition has progressed significantly in the last few years, there are still many open challenges. In this section, we first consider challenges specific to a particular sensor modality and then we look at challenges which are generalizable to all transfer-based activity recognition.
As can be seen in Table 5, performing transfer-based activity recognition when the source data is not labeled has not received much attention in current research. Outside the domain of activity recognition, researchers have leveraged the unlabeled source data to improve transfer in the target domain (Dai, Yang, Xue and Yu, 2008; Raina, Battle, Lee, Packer and Ng, 2007; Wang, Song and Zhang, 2008) but such techniques have yet to be applied to activity recognition.
Table 5.
Another area needing more attention is relational-knowledge transfer for activity recognition as indicated in Table 6. Relational-knowledge transfer requires that there exist certain relationships in the data which can be learned and transfered across populations. Data for activity recognition has the potential to contain such transferable relationships indicating that this may be an important technique to pursue. See (Davis and Domingos, 2009; Mihalkova, Huynh and Mooney, 2007; Mihalkova and Mooney, 2008; Mihalkova and Mooney, 2009) for examples of relational-knowledge transfer.
Table 6.
Tables 4–6 also indicate several more niche areas which could be further investigated. For example, in the video camera domain, most of the work has focused on informed supervised parameter-based transfer learning, while the other techniques have not been heavily applied. Similarly, transferring across different labels spaces is a much less studied problem in transfer-based activity recognition. Finally, we note that parameter-based transfer learning is also less studied for the ambient sensor modality.
Table 4.
The current direction of most transfer-based activity recognition is to push the limits on how different the source and target domains and tasks can be. The scenarios discussed in Section 2 illustrate the importance of continuing in this direction. More work is needed to improve transfer across sensor modalities and to transfer knowledge across multiple differences. Instead of transferring learning from one smart home environment to another, can we transfer from a smart home to a smart workplace or smart hospital? We envision one day chaining multiple transfers together to achieve even greater diversity between the source and target populations.
As researchers continue to expand the applicability of transfer learning, two natural questions arise. First, can we define a generalizable distance metric for determining the difference between the source and target populations? Some domain-specific distances have been used in the past, but it would be useful if we had a domain-independent distance measure. This measure could be used to facilitate comparisons between different transfer learning approaches as well as provide an indication of whether transfer learning should even be applied in a given situations. Such a measure would need to indicate how the source and target data differ (feature-space, marginal probabilities, label-space, and, objective predictive function) as well as quantify the magnitude of the differences. Second, can we detect and prevent the occurrence of negative transfer effects. Negative transfer effects occur when the use of transfer learning actually decreases performance instead of increasing performance. These two questions are actually related, because an accurate distance metric may provide an indication of when negative transfer will occur for a given transfer learning technique. Rosenstein et al. look at the question of when to use transfer learning in (Rosenstein, Marx, Kaelbling and Dietterich, 2005). They empirically show that when two tasks are of sufficient dissimilarity, negative transfer occurs. Mahmud and Ray define a distance metric for measuring the similarity of two tasks based on the conditional Kolmogorov complexity between the tasks and prove some theoretical bounds using this distance measure (Mahmud and Ray, 2008).
This survey has reviewed the current literature regarding transfer based activity recognition. We discussed several promising techniques and consider the many open challenges that still need to be addressed in order to progress the field of transfer learning for activity recognition.
Biographies
Diane Cook is a Huie-Rogers Chair Professor in the School of Electrical Engineering and Computer Science at Washington State University. Dr. Cook received a B.S. degree in Math/Computer Science from Wheaton College in 1985, a M.S. degree in Computer Science from the University of Illinois in 1987, and a Ph.D. degree in Computer Science from the University of Illinois in 1990. Her research interests include artificial intelligence, machine learning, graph-based relational data mining, smart environments, and robotics. Dr. Cook is an IEEE Fellow.
Kyle D. Feuz is an IGERT Fellow in the School of Electrical Engineering and Computer Science at Washington State University where he is working towards the completion of his PhD in Computer Science. He received a B.S. degree in Computer Science from Utah State University in 2010 and a M.S degree in Computer Science from Utah State University in 2011. His research interests are in the areas of machine learning, activity recognition, multi-agent systems and human-computer interaction.
Narayanan C. Krishnan Narayanan C Krishnan completed his PhD in Computer Science in December 2010 from Arizona State University. He is currently working as Assistant Research Professor at Washington State University. Narayanan received his Bachelors and Masters in Science majoring in Mathematics from Sri Sathya Sai Institute of Higher Learning in 2000 and 2002 respectively. He then went on to complete his Masters in Technology (Computer Science) also from the same university in 2004. His research interests are in the area of activity recognition, pattern recognition and machine learning for pervasive computing applications.
References
- Agrawal R, Srikant R. Mining sequential patterns. Proceedings of the International Conference on Data Engineering; 1995. pp. 3–14. [Google Scholar]
- Alemdar H, Ersoy C. Wireless sensor networks for healthcare: A survey. Computer Networks. 2010;54(15):2688–2710. http://www.sciencedirect.com/science/article/pii/S1389128610001398. [Google Scholar]
- Arnold A, Nallapati R, Cohen W. A comparative study of methods for transductive transfer learning. Data Mining Workshops, 2007. ICDM Workshops 2007. Seventh IEEE International Conference on; 2007. pp. 77–82. [Google Scholar]
- Avci A, Bosch S, Marin-Perianu M, Marin-Perianu R, Havinga P. Activity recognition using inertial sensing for healthcare, wellbeing and sports applications: A survey. Architecture of Computing Systems (ARCS), 2010 23rd International Conference on; 2010. pp. 1–10. [Google Scholar]
- Barnett S, Ceci S. When and where do we apply what we learn?: A taxonomy for far transfer. Psychological bulletin. 2002;128(4):612–637. doi: 10.1037/0033-2909.128.4.612. [DOI] [PubMed] [Google Scholar]
- Blanke U, Schiele B. Remember and transfer what you have learned-recognizing composite activities based on activity spotting. Wearable Computers (ISWC), 2010 International Symposium on; IEEE. 2010. pp. 1–8. [Google Scholar]
- Byrnes J. Cognitive development and learning in instructional contexts. Allyn and Bacon; Boston: 1996. [Google Scholar]
- Calatroni A, Roggen D, Tröster G. Automatic transfer of activity recognition capabilities between body-worn motion sensors: Training newcomers to recognize locomotion. Eighth International Conference on Networked Sensing Systems (INSS’11); Penghu, Taiwan. 2011. [Google Scholar]
- Cao L, Liu Z, Huang T. Cross-dataset action detection. Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on; 2010. pp. 1998–2005. [Google Scholar]
- Chan M, Estve D, Escriba C, Campo E. A review of smart homespresent state and future challenges. Computer Methods and Programs in Biomedicine. 2008;91(1):55–81. doi: 10.1016/j.cmpb.2008.02.001. [DOI] [PubMed] [Google Scholar]
- Chang C-C, Lin C-J. Libsvm: A library for support vector machines. ACM Transactions on Intelligent Systems and Technology. 2011;2(3):27:1–27:27. [Google Scholar]
- Chattopadhyay R, Krishnan N, Panchanathan S. Topology preserving domain adaptation for addressing subject based variability in semg signal. 2011 AAAI Spring Symposium Series.2011. [Google Scholar]
- Chieu H, Lee W, Kaelbling L. Technical report. Singapore-MIT Alliance (SMA); 2006. Activity recognition from physiological data using conditional random fields. [Google Scholar]
- Cook D. Intelligent Systems. 99. IEEE; 2010. Learning setting-generalized activity models for smart spaces; p. 1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dai W, Yang Q, Xue G-R, Yu Y. Boosting for transfer learning. Proceedings of the 24th international conference on Machine learning; New York, NY, USA: ACM; 2007. pp. 193–200. ICML ’07. [Google Scholar]
- Dai W, Yang Q, Xue G-R, Yu Y. Self-taught clustering. Proceedings of the 25th international conference on Machine learning; New York, NY, USA: ACM; 2008. pp. 200–207. ICML ’08. [Google Scholar]
- Davis J, Domingos P. Deep transfer via second-order markov logic. Proceedings of the 26th Annual International Conference on Machine Learning; New York, NY, USA: ACM; 2009. pp. 217–224. ICML ’09. [Google Scholar]
- Duan L, Xu D, Tsang I, Luo J. Visual event recognition in videos by learning from web data. Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on; 2010. pp. 1959–1966. [DOI] [PubMed] [Google Scholar]
- Elkan C. The foundations of cost-sensitive learning. Proceedings of the 17th international joint conference on Artificial intelligence - Volume 2; San Francisco, CA, USA: Morgan Kaufmann Publishers Inc; 2001. pp. 973–978. IJCAI’01. [Google Scholar]
- Farhadi A, Tabrizi M. Learning to recognize activities from the wrong view point. In: Forsyth D, Torr P, Zisserman A, editors. Computer Vision ECCV 2008. Springer; Berlin/Heidelberg: 2008. pp. 154–166. Vol. 5302 of Lecture Notes in Computer Science. [Google Scholar]
- Freund Y, Schapire RE. A decision-theoretic generalization of on-line learning and an application to boosting. Journal of Computer and System Sciences. 1997;55(1):119–139. [Google Scholar]
- Gu T, Chen S, Tao X, Lu J. An unsupervised approach to activity recognition and segmentation based on object-use fingerprints. Data and Knowledge Engineering. 2010;69(6):533–544. [Google Scholar]
- Hachiya H, Sugiyama M, Ueda N. Importance-weighted least-squares probabilistic classifier for covariate shift adaptation with application to human activity recognition. Neurocomputing. 2012;80(0):93–101. Special Issue on Machine Learning for Signal Processing 2010. [Google Scholar]
- Haigh K, Yanco H. Automation as caregiver: A survey of issues and technologies. AAAI-02 Workshop on Automation as Caregiver: The Role of Intelligent Technology in Elder Care; 2002. pp. 39–53. [Google Scholar]
- Hu D, Yang Q. Transfer learning for activity recognition via sensor mapping. Twenty-Second International Joint Conference on Artificial Intelligence.2011. [Google Scholar]
- Hu D, Zheng V, Yang Q. Cross-domain activity recognition via transfer learning. Pervasive and Mobile Computing. 2010;7(3):344–358. [Google Scholar]
- Kasteren TL, Englebienne G, Kröse BJ. An activity monitoring system for elderly care using generative and discriminative models. Personal and Ubiquitous Computing. 2010;14(6):489–498. [Google Scholar]
- Kim E, Helal S, Cook D. Human activity recognition and pattern discovery. Pervasive Computing, IEEE. 2010;9(1):48–53. doi: 10.1109/MPRV.2010.7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Krishnan N. PhD thesis. Arizona State University; 2010. A Computational Framework for Wearable Accelerometer-Based. [Google Scholar]
- Krishnan N, Lade P, Panchanathan S. Activity gesture spotting using a threshold model based on adaptive boosting. Multimedia and Expo (ICME), 2010 IEEE International Conference on; 2010. pp. 155–160. [Google Scholar]
- Krishnan N, Panchanathan S. Analysis of low resolution accelerometer data for continuous human activity recognition. Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on; 2008. pp. 3337–3340. [Google Scholar]
- Kurz M, Hölzl G, Ferscha A, Calatroni A, Roggen D, Tröster G. Real-time transfer and evaluation of activity recognition capabilities in an opportunistic system. ADAPTIVE 2011, The Third International Conference on Adaptive and Self-Adaptive Systems and Applications; 2011. pp. 73–78. [Google Scholar]
- Kwapisz JR, Weiss GM, Moore SA. Activity recognition using cell phone accelerometers. Proceedings of the Fourth International Workshop on Knowledge Discovery from Sensor Data; 2010. pp. 10–18. [Google Scholar]
- Lam A, Roy-Chowdhury A, Shelton C. Interactive event search through transfer learning. In: Kimmel R, Klette R, Sugimoto A, editors. Computer Vision - ACCV 2010. Springer; Berlin/Heidelberg: 2011. pp. 157–170. Vol. 6494 of Lecture Notes in Computer Science. [Google Scholar]
- Lester J, Choudhury T, Kern N, Borriello G, Hannaford B. A hybrid discriminative/generative approach for modeling human activities. Proceedings of the 19th international joint conference on Artificial intelligence; San Francisco, CA, USA: Morgan Kaufmann Publishers Inc; 2005. pp. 766–772. [Google Scholar]
- Liao L, Fox D, Kautz H. Location-based activity recognition using relational markov networks. Proceedings of the 19th International Joint Conference on Artificial Intelligence; San Francisco, CA, USA: Morgan Kaufmann Publishers Inc; 2005. pp. 773–778. [Google Scholar]
- Liu J, Shah M, Kuipers B, Savarese S. Cross-view action recognition via view knowledge transfer. Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on; 2011. pp. 3209–3216. [Google Scholar]
- Logan B, Healey J, Philipose M, Tapia EM, Intille S. A long-term evaluation of sensing modalities for activity recognition. Proceedings of the 9th international conference on Ubiquitous computing; Berlin, Heidelberg: Springer-Verlag; 2007. pp. 483–500. [Google Scholar]
- Mahmud MM, Ray S. Transfer learning using kolmogorov complexity: Basic theory and empirical evaluations. In: Platt J, Koller D, Singer Y, Roweis S, editors. Advances in Neural Information Processing Systems 20. MIT Press; Cambridge, MA: 2008. pp. 985–992. [Google Scholar]
- Maurer U, Smailagic A, Siewiorek D, Deisher M. Activity recognition and monitoring using multiple sensors on different body positions. International Workshop on Wearable and Implantable Body Sensor Networks.2006. [Google Scholar]
- Mihalkova L, Huynh T, Mooney R. Mapping and revising markov logic networks for transfer learning. Proceedings of the national conference on artificial intelligence; Menlo Park, CA; Cambridge, MA; London. AAAI Press; MIT Press; 2007. p. 608. 1999. [Google Scholar]
- Mihalkova L, Mooney R. Transfer learning by mapping with minimal target data. Proceedings of the AAAI-08 Workshop on Transfer Learning for Complex Tasks.2008. [Google Scholar]
- Mihalkova L, Mooney RJ. Transfer learning from minimal target data by mapping across relational domains. Proceedings of the 21st international jont conference on Artifical intelligence; San Francisco, CA, USA: Morgan Kaufmann Publishers Inc; 2009. pp. 1163–1168. IJCAI’09. [Google Scholar]
- Nater F, Tommasi T, Grabner H, van Gool L, Caputo B. ICCV WS on Visual Surveillance. 2011. Transferring activities: Updating human behavior analysis (both first authors contributed equally) [Google Scholar]
- Palmes P, Pung HK, Gu T, Xue W, Chen S. Object relevance weight pattern mining for activity recognition and segmentation. Pervasive and Mobile Computing. 2010;6(1):43–57. [Google Scholar]
- Pan J, Yang Q, Chang H, Yeung D. A manifold regularization approach to calibration reduction for sensor-network based tracking. Proceedings of the National Conference on Artificial Intelligence; 2006. p. 988. [Google Scholar]
- Pan SJ, Tsang IW, Kwok JT, Yang Q. Domain adaptation via transfer component analysis. IEEE Transactions on Neural Networks. 2011;22(2):199–210. doi: 10.1109/TNN.2010.2091281. [DOI] [PubMed] [Google Scholar]
- Pan S, Kwok J, Yang Q, Pan J. Adaptive localization in a dynamic wifi environment through multi-view learning. Proceedings of the National Conference on Artificial Intelligence; 2007. p. 1108. [Google Scholar]
- Pan S, Shen D, Yang Q, Kwok J. Transferring localization models across space. Proceedings of the 23rd national conference on Artificial intelligence; 2008. pp. 1383–1388. [Google Scholar]
- Pan S, Yang Q. A survey on transfer learning. Knowledge and Data Engineering, IEEE Transactions on. 2010;22(10):1345–1359. [Google Scholar]
- Pan S, Zheng V, Yang Q, Hu D. Transfer learning for wifi-based indoor localization. Association for the Advancement of Artificial Intelligence (AAAI) Workshop; 2008. p. 6. [Google Scholar]
- Philipose MPK, Perkowitz M, Patterson DJ, Fox D, Kautz H, Hhnel D. Inferring activities from interactions with objects. IEEE Pervasive Computing. 2004;3:50–57. [Google Scholar]
- Raina R, Battle A, Lee H, Packer B, Ng AY. Self-taught learning: transfer learning from unlabeled data. Proceedings of the 24th international conference on Machine learning; New York, NY, USA: ACM; 2007. pp. 759–766. ICML ’07. [Google Scholar]
- Rashidi P, Cook D. Transferring learned activities in smart environments. 5th International Conference on Intelligent Environments; 2009. pp. 185–192. [Google Scholar]
- Rashidi P, Cook D. Activity recognition based on home to home transfer learning. AAAI Workshop on Plan, Activity, and Intent Recognition.2010a. [Google Scholar]
- Rashidi P, Cook D. Multi home transfer learning for resident activity discovery and recognition; KDD Knowledge Discovery from Sensor Data; 2010b. pp. 56–63. [Google Scholar]
- Rashidi P, Cook D. Activity knowledge transfer in smart environments. Pervasive and Mobile Computing. 2011;7(3):331–343. [Google Scholar]
- Rashidi P, Cook D, Holder L, Schmitter-Edgecombe M. Discovering activities to recognize and track in a smart environment. IEEE Transactions on Knowledge and Data Engineering. 2011;23(4):527–539. doi: 10.1109/TKDE.2010.148. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Roggen D, Frster K, Calatroni A, Trster G. The adarc pattern analysis architecture for adaptive human activity recognition systems. Journal of Ambient Intelligence and Humanized Computing Online. 2011:1–18. doi: 10.1007/s12652-011-0064-0. [DOI] [Google Scholar]
- Rosenstein MT, Marx Z, Kaelbling LP, Dietterich TG. To transfer or not to transfer. In NIPS05 Workshop, Inductive Transfer: 10 Years Later.2005. [Google Scholar]
- Taylor M, Stone P. Transfer learning for reinforcement learning domains: A survey. The Journal of Machine Learning Research. 2009;10:1633–1685. [Google Scholar]
- Tenenbaum JB, Silva Vd, Langford JC. A global geometric framework for nonlinear dimensionality reduction. Science. 2000;290(5500):2319–2323. doi: 10.1126/science.290.5500.2319. [DOI] [PubMed] [Google Scholar]
- Thorndike E, Woodworth R. The influence of improvement in one mental function upon the efficiency of other functions.(i) Psychological review. 1901;8(3):247–261. [Google Scholar]
- Thrun S. Explanation-based neural network learning: A lifelong learning approach. Kluwer Academic Publishers; 1996. [Google Scholar]
- Thrun S, Pratt L. Learning to learn. Kluwer Academic Publishers; 1998. [Google Scholar]
- van Kasteren T, Englebienne G, Kröse B. Recognizing activities in multiple contexts using transfer learning. AAAI AI in Eldercare Symposium.2008. [Google Scholar]
- van Kasteren T, Englebienne G, Krse B. Transferring knowledge of activity recognition across sensor networks. In: Floren P, Krger A, Spasojevic M, editors. Pervasive Computing. Springer; Berlin/Heidelberg: 2010. pp. 283–300. Vol. 6030 of Lecture Notes in Computer Science. [Google Scholar]
- Venkatesan A. PhD thesis. Arizona State University; 2011. A Study of Boosting based Transfer Learning for Activity and Gesture Recognition. [Google Scholar]
- Venkatesan A, Krishnan N, Panchanathan S. Cost-sensitive boosting for concept drift. International Workshop on Handling Concept Drift in Adaptive Information Systems; 2010; 2010. pp. 41–47. [Google Scholar]
- Vilalta R, Drissi Y. A perspective view and survey of meta-learning. Artificial Intelligence Review. 2002;18:77–95. [Google Scholar]
- Wang Z, Song Y, Zhang C. Transferred dimensionality reduction. In: Daelemans W, Goethals B, Morik K, editors. Machine Learning and Knowledge Discovery in Databases. Springer; Berlin/Heidelberg: 2008. pp. 550–565. Vol. 5212 of Lecture Notes in Computer Science. [Google Scholar]
- Wei B, Pal C. Heterogeneous transfer learning with rbms. Twenty-Fifth AAAI Conference on Artificial Intelligence.2011. [Google Scholar]
- Wu C, Khalili AH, Aghajan H. Multiview activity recognition in smart homes with spatio-temporal features. Proceedings of the Fourth ACM/IEEE International Conference on Distributed Smart Cameras; New York, NY, USA: ACM; 2010. pp. 142–149. ICDSC ’10. [Google Scholar]
- Xian-ming L, Shao-zi L. Transfer adaboost learning for action recognition. IT in Medicine Education, 2009. ITIME ’09. IEEE International Symposium on; 2009. pp. 659–664. [Google Scholar]
- Yang J, Yan R, Hauptmann AG. Cross-domain video concept detection using adaptive svms. Proceedings of the 15th international conference on Multimedia; New York, NY, USA: ACM; 2007. pp. 188–197. MULTIMEDIA ’07. [Google Scholar]
- Yang Q. Activity recognition: linking low-level sensors to high-level intelligence. Proceedings of the 21st international jont conference on Artifical intelligence; Morgan Kaufmann Publishers Inc; 2009. pp. 20–25. [Google Scholar]
- Yang W, Wang Y, Mori G. Learning transferable distance functions for human action recognition. In: Wang L, Zhao G, Cheng L, Pietikinen M, editors. Machine Learning for Vision-Based Motion Analysis. Springer; London: 2011. pp. 349–370. Advances in Pattern Recognition. [Google Scholar]
- Zhao Z, Chen Y, Liu J, Liu M. Cross-mobile elm based activity recognition. International Journal of Engineering and Industries. 2010;1(1):30–38. [Google Scholar]
- Zhao Z, Chen Y, Liu J, Shen Z, Liu M. Cross-people mobile-phone based activity recognition. Twenty-Second International Joint Conference on Artificial Intelligence.2011. [Google Scholar]
- Zheng V, Hu D, Yang Q. Cross-domain activity recognition. Ubicomp. 2009;9:61–70. [Google Scholar]
- Zheng V, Pan S, Yang Q, Pan J. Transferring multi-device localization models using latent multi-task learning. Proceedings of the 23rd national conference on Artificial intelligence; 2008. pp. 1427–1432. [Google Scholar]