Abstract
Advances in wearable sensing and mobile computing have enabled the collection of health and well-being data outside of traditional laboratory and hospital settings, paving the way for a new era of mobile health. Meanwhile, artificial intelligence (AI) has made significant strides in various domains, demonstrating its potential to revolutionize healthcare. Devices can now diagnose diseases, predict heart irregularities and unlock the full potential of human cognition. However, the application of machine learning (ML) to mobile health sensing poses unique challenges due to noisy sensor measurements, high-dimensional data, sparse and irregular time series, heterogeneity in data, privacy concerns and resource constraints. Despite the recognition of the value of mobile sensing, leveraging these datasets has lagged behind other areas of ML. Furthermore, obtaining quality annotations and ground truth for such data is often expensive or impractical. While recent large-scale longitudinal studies have shown promise in leveraging wearable sensor data for health monitoring and prediction, they also introduce new challenges for data modelling. This paper explores the challenges and opportunities of human-centred AI for mobile health, focusing on key sensing modalities such as audio, location and activity tracking. We discuss the limitations of current approaches and propose potential solutions.
Keywords: mobile health, wearable sensing, artificial intelligence for health
1. Introduction
Advances in the ways to collect and process health monitoring data can transform our society from multiple dimensions such as better access to preventive healthcare, low-cost diagnosis and long-term monitoring. An overwhelming majority of today’s medical services collect such data in laboratory settings. While an average person visits a doctor only around 5 times a year [1] these measurements provide sparse snapshots of one’s health. On the other hand, recent advances in wearable sensing and mobile computing, along with their wide adoption, have created new pathways for the collection of health and well-being data outside of laboratory and hospital settings, in a longitudinal fashion. Apart from filling the gaps of traditional clinical data, these devices open up new research and commercial directions for large-scale lifestyle monitoring. For example, millions of people worldwide use personal devices to track their physical activity and sleep, with increasingly more sophisticated predictive capabilities [2]. The penetration of mobile devices has introduced scale: many more individuals can now be reached and assessed. For example, in a hospital environment, mobile experience sampling enabled the collection of 11 381 survey responses over a 12-month period from 304 physicians and nurses, completed with minimal initial training [3]. Mobile sensors enable researchers to collect not only the explicit reports of the participants, but also the context in which these answers were provided. Indeed, a survey of 110 experience-sampling papers concluded that a total of 70 studies (63.6%) passively or actively collected sensor data from the participants’ study device [4].
At the same time, artificial intelligence (AI) has taken a huge step forward across domains ranging from object recognition in images [5], to winning against the best players in the games of Go, Atari and chess [6], or outperforming experts in breast cancer screening [7]. The widespread popularity and dramatic improvements in predictive modelling came from mature open-source scientific software libraries, easier data crowdsourcing and labelling, and the re-purposing of specialized hardware like graphics cards. In this article, we discuss how the adaptation of machine learning (ML)1 to mobile sensing-based human health applications is moving towards a new era of mobile health.
Sensors in wearable and mobile devices such as accelerometers, electrocardiograms (ECGs), global positioning systems (GPSs), gyroscopes and microphones monitor our location, sleep, steps, eating and working habits in the real world. Their continuous tracking produces huge datasets with the potential to create a holistic understanding of the most important components of our everyday health [8]. Results from a variety of fields, for example, physical activity [9], disease diagnosis [10] and mental health disorders [11], have consistently shown solid evidence that mobile health can be potentially employed for human health monitoring in the future.
Although the value of mobile sensing2 has been recognized, it has not been kept up with other application areas of machine learning. For instance, over the last decade devices such as Fitbit or the iPhone have been collecting multi-modal sensor data at an unprecedented temporal resolution. However, effectively leveraging these datasets presents many challenges. Some of the obstacles lie in (i) the noisy measurements of the sensor data which are hard to model due to their correlation over time or their nonlinear structure [13]; (ii) the high-dimensional raw data or engineered feature representations that are difficult to handle using conventional AI algorithms [14]; (iii) sparse and irregular-sampled time-series data; (iv) heterogeneity in different data types [15], etc. These all lead to these data being frequently overlooked for scientific and medical research. Further, obtaining quality annotations and ground truth is costly or even impossible at this granularity. The most established research and clinical practice for annotation is still based on pen-and-paper self-reports and surveys which are time-consuming and also incomplete and scarce [16].
While mobile health research data traditionally have a small number of participants and limited-sized measurements, recently there have been several large longitudinal studies like the UK Biobank [17], the Apple Study [18], the Fenland Study [19], Utsureko [20] and EmotionSense [21] with tens of thousands of participants using wearable sensors. These large studies have started showing their potential too; for example, an elevated resting heart rate (RHR) from over 200 000 Fitbit users was used to predict influenza-like illness in the USA [22]. However, these large studies introduced new challenges for data modelling. Statistical methods developed for small datasets often fail in these cases [23]. Additionally, how to understand and effectively model the longitudinal data over time to achieve reliable monitoring and forecasting of individuals’ health status is of great potential for more flexible and convenient health delivery. Despite these successes witnessed in mobile health, there are also inevitably a multitude of implications and limitations existing due to modelling, such as privacy concerns, human-introduced biases, practical deployment in clinical practice, and improved usability and user experience for human–computer interfaces.
This paper discusses the challenges and opportunities of human-centred AI for mobile health, through case studies in some key mobile sensing modalities such as audio, location and activity tracking, as shown in figure 1. We particularly focus on these areas considering their popularity within both research and commercial applications. Additionally, these three modalities rely on sensors commonly found in the majority of mobile and wearable devices, unlike biosignal collection devices such as ECG chest straps or EEG headsets. Moreover, they typically involve passive sensing, minimizing the need for active user participation as opposed to camera-based vision techniques. These advantages have led to extensive research in these three modalities [24–26], suggesting their potential for future large-scale population screening and monitoring. However, we acknowledge other emerging areas such as radio sensing or biosignals in §2.5. Further, we explore the underlying limitations of current approaches, which are common across various modalities, and delve into potential solutions for these challenges.
2. AI for mobile health—case studies
Data analysis techniques typically are specific to the sensing modality and application at hand. However, they all follow the same general structure involving three main stages: data pre-processing, feature extraction and modelling. The pre-processing step can involve data cleaning, transformation and augmentation. The pre-processing also involves segmenting the continuous stream of data according to the specific application. The segments can simply be the sliding windows on the data stream or it can be more complex to describe and extract, for example, cough sounds detected and extracted from long noisy audio streams [27] prior to cough based modelling or trips extracted from continuous GPS traces [28]. Features are then extracted from the segments. Both human designed and learnt features have been explored and we discuss them separately for different modalities. If multiple modalities are analysed in a study, feature representations are fused [15]. ML techniques are then used for model development [29,30]. Multi-modal fusion can also be carried out at this stage, generally by fusing model outputs from different modalities [31].
Let us now take a deep dive into the three most popular sensing modalities in mobile health studies, namely, audio, location and motion sensors, and a brief look at the other modalities, including Bluetooth, Wi-Fi, etc. For each modality we start with its popular applications showing its potential to be used in mobile health applications, followed by a discussion of the common ways it is recorded, and finally detailing its mobile health applications.
2.1. Audio
Audio sensing at scale. Audio, as one of the most natural human communication signals, contains rich information. The speech and audio processing community has explored the potential of audio data for a wide range of industry applications, e.g. virtual assistants with automatic speech recognition [32], biometric system safety with speaker verification or voice antispoofing [33], etc. Recently, audio has also been analysed for health problems, including both mental and physical health, such as emotion detection [34–36], depression detection [37–39], Alzheimer’s and dementia detection [40–42], etc. Despite its wide application and great potential, most of them were carried out using data collected in the laboratory or a controlled environment. Only limited work has explored the potential of audio within the mobile health context in free-living environment. With the rapid development of mobile technologies such as smartphones, tablets and wearables, it allows for greater flexibility in data collection, processing and transmitting, offering the potential for healthcare support, delivery and intervention in a more flexible, scalable and prompt way.
2.1.1. Data
Data types. An overview of audio signals for health-related applications is shown in figure 2. A variety of different types of audio sounds can be easily collected from mobile phones and wearbles (e.g. smartwatch, earbuds) in daily life. Speech, as one of the dominant sounds, contains rich information that can infer human mental health and physical health [34,39]. Breathing and cough are two types of sounds that contain clinically relevant information and highly related to pulmonary health [10,43]. In addition, other types of sounds (e.g. eating sounds recorded via microphones) can also be informative for obesity health.
In addition to the normal airborne sound, bone conducted sounds collected via wearables have also been analysed for health application, which is generally collected via earables or sensors placed near neck. Chen et al. [44] collected sound signals from a small acoustic sensor placed at the neck to estimate heart rate (HR) and conduct heart sound classification. One other study [45] has investigated acoustics in the ear canal, measuring in-ear pulse wave velocity using heart signal as reference and the other detecting respiratory sounds [46]. Amft et al. [47] collected chewing sounds acquired with a microphone located inside the ear canal and used it to classify different kinds of food. Some example datasets can be found in table 1.
Table 1.
dataset | samples | participants | data types | labels | |
---|---|---|---|---|---|
audio | Coswara [48] | 2030 | — | breathing, cough, speech | COVID-19, smoking status |
COVID-19 sounds [49] | 53 449 | 36 116 | breathing, cough, speech | COVID-19, smoking status, language | |
RECOLA [50] | 34 | 34 | speech | arousal, valence (emotion) | |
E-DAIC [39] | — | 275 | speech | PHQ-8 (depression) | |
ADReSS challenge [40] | 4076 | 156 | voice/speech | Alzheimer’s versus non-AD, MMSE score | |
chewing events [51]a | — | 5 | chewing sound | eating activities | |
mobility | COVID-19 Community Mobility Reports [52] | — | aggregated | aggregated data on the change in mobility trends during COVID-19 | — |
mobile phone mobility data [53] | — | aggregated | aggregated movement between regions during COVID-19 | — | |
motion | GLOBEM dataset: multi-year datasets for longitudinal human behaviour modelling generalization [54] | — | 497 (700 user-years) | activity (steps, sleep), location, call logs, bluetooth, screen | depression, personality, well-being and more |
Multi-Ethnic Study of Atherosclerosis (MESA) [55] | 2 200 000 | 1817 | polysomnography (PSG), actigraphy (activity) | sleep–wake classification | |
FitRec [56] | 253 020 | 1104 | heart rate, speed, GPS | workout route prediction, heart rate prediction |
aData for chewing events are not publicly available.
Clinical relevance. It was proven that changes in affective state and cognitive effects associated with depression have altered the mechanisms involved in speech production [57,58]. Lung disease is also expected to alter the sound generation mechanism. These changes will provide valuable information to identify between the healthy and non-healthy conditions. Some studies have analysed the significance and correlation of different audio features with mental disorder or physical disease. Though they may not be directly estimated using data collected using mobile devices, the findings should reveal certain information. Acoustic properties of speech in terms of formants and power distribution were found to be best discriminators to distinguish between depression and suicidality individuals and healthy individuals [59]. Many other studies also obtained the similar findings, and details can be found in a review paper [60]. Voice biomarkers have also been analysed for recovery from acute respiratory illness [61]. The findings suggest that jitter, shimmer and spectral contrast within certain frequency bands show consistent trends for recovering participants, and can be informative to distinguish them from healthy users as well. Slightly different conclusions might be made for different studies due to the data variability and data size, but the general conclusion of the clinical relevance of audio sounds to health and the distinguishable information between still holds.
2.1.2. Modelling
Statistical analysis. Prior to model development for mobile health applications, effective sound data types or data quality needs to be guaranteed. Dubey et al. [62] studied speech data collected from a smartwatch’s built-in microphone in a controlled environment, and targeted on remotely monitoring of PD patients’ speech and voices exercises to improve the communication effectiveness of speech therapy outside the clinic. The study analysed and compared speech characteristics in terms of the loudness and fundamental frequency to that collected using traditional recording methods, and showed comparable results demonstrating the quality of audio data collected through smartwatches. A continuous mobile sensing app is developed to monitor students’ life on stress, sleep, activity, mood, sociability, mental well-being and academic performance. Audio data, as one of the modalities, are recorded using smartphones. Significant correlations are obtained between conversation duration and/or frequency computed from audio data and PHQ-9 depression scores, flourishing scale and perceived stress scale (PSS) [63]. One other study [64] also analysed the audio quality of speech and respiratory sounds (e.g. cough, sneezing, etc.) collected from smartwatches in the wild, and demonstrated that the quality is good enough for humans to be able to detect speech and respiratory sounds. It also suggests the difficulty in extracting these sounds automatically from collected audio due to a wide variety of background noise. A multi-modal sensor platform for knee joint health assessment is also developed, which includes the joint sounds collected through multi-microphones [65]. The significant acoustic emissions were found to occur at similar joint angles for repeated exercises, suggesting the reliability of the audio sounds for knee activities. These statistical analysis or proof-of-concept results demonstrate the potential of audio sounds collected through different mobile devices for health monitoring.
ML techniques. Different ML techniques ranging from traditional methods such as support vector machines (SVMs) to deep learning techniques have been employed. Many studies using wearables such as smartwatches and earables only involve limited number of participants, where conventional ML techniques are suitable to avoid overfitting. Logistic Regression, Decision Tree and Random Forest have been employed for chewing sounds detection [47,51], eating behaviour monitoring [66], stress detection [67], etc. SVM is also a competitive technique for classification, and has been validated for stress detection [67], sleep event detection [30], cognitive impairment detection [68].
With more flexible data collection using smartphones and cloud servers, many smartphone-based studies collect audio data on a relatively large scale. Deep learning techniques have been widely employed as they generally show superior performance over traditional ML techniques with adequate and high-quality data. Different neural network structures have been validated for COVID-19 detection. A convolutional neural network (CNN) based network structure is validated using a combination of crowdsourced voice, cough and breathing sounds [10,49]. The system is developed and validated using a total of 2478 participants and yields 0.71 in terms of ROC-AUC, suggesting the potential of crowdsourced audio for COVID-19 detection. A residual network (ResNet) [29] is also employed for COVID-19 detection using cough and breathing sounds. Long short-term memory (LSTM) [69], gated recurrent units (GRUs) [70] and neural ODE processes [71] are also beneficial for COVID-19 detection and especially for COVID-19 progression prediction and forecasting, as they capture the temporal information over time in either the short term or the long term. Variations on the network structures or training strategies have also been investigated, such as the attention mechanism [72] and self-supervised learning schemes [73]. Deep learning has also been employed for lung health detection. A pre-trained CNN based network of MobileNet is employed to detect respiratory disease, and achieves an F1 score of 0.912 on healthy versus unhealthy, 0.870 on obstructive versus non-obstructive and 0.813 on COPD versus asthma classification [74]. Two types of lung sounds, namely, wheeze and crackle, are detected by semi-supervised deep learning [75].
2.1.3. Applications
Audio sounds contain rich health-related information and different types of audio sounds additionally reveal complementary information. It can be used for both mental and physical health detection and monitoring.
Mental health is important in daily life and affects a variety of human behaviours including thoughts, mood and actions. The increasing burden of mental disorders significantly impacts health and major social, human rights and economic consequences in all countries of the world. Automatic detection and monitoring of mental health could provide prompt diagnosis and reduce the burden. Depression, as one of the leading mental disorders, has been studied using audio sounds [76]. Other mental disorders have also been investigated using audio sensing, including bipolar [66], schizophrenia [77], autism [78]. Stress and mood which are highly correlated to mental health have also been studied using audio sounds [67,76].
Physical health studies were mainly carried out for respiratory disease and sleep health. COVID-19 has been studied using different audio sounds, including voice, breathing and cough sounds [10,29,79,80]. Breathing patterns are extracted from audio signal to monitor sleep apnoea [81], and another study explored the potential of breathing rate by leveraging smartphones for sleep monitoring, which detects sleep events including snore, cough, turn over and get up [30]. Some studies used the microphone of smartphones to quantify the flow and volume of air inspired and expired by the lungs during breathing, which reflects the degree of airflow obstruction in clinically associated pulmonary disorders such as COPD and asthma [82,83]. Heart rate and breathing rate are also estimated from in-ear microphones, and demonstrated the potential of earables for health monitoring [44,84]. This suggests new pathways for cardiovascular health monitoring using wearables. Eating behaviours are also analysed using audio collected from smartwatches and earables [47,51,85]. While audio-based applications have seen significant advancements, deploying them in real-world settings, especially for audio-based COVID-19 detection, remains challenging [86]. This primarily stems from difficulties acquiring high-quality data during the pandemic, managing variations in recording environments and devices, accounting for symptom differences associated with various virus strains, which can cause distinct audio variations, etc.
2.2. Mobility
Localization sensors are ubiquitous in mobile phones and wearable devices. They can measure user mobility in different spatial to temporal resolutions ranging from continuous tracking using GPS to intermittent recording of connected cell tower locations. Tracking locations and analysing mobility has contributed to many day-to-day applications, for example, traffic management and forecasting [87,88], routing individuals [89] and delivery packages [90], recommender systems [91], dynamic load balancing systems for cellular networks [92], to name a few. These applications have used techniques ranging from statistics to deep learning, and more theoretical computational geometry and algebraic topology.
Mobility naturally captures the lifestyle of individuals and thus related health information; for example, the type of restaurant a person visits reveals food habit and hence correlated health problems. In spite of its potential to unlock secrets of human health conditions, mobility has seen limited applications for mobile health primarily due to lack of large relevant datasets. Here, we first discuss the types of mobility data followed by popular features and finally health applications.
2.2.1. Data types
Location of a device can be captured in various ways differing in spatial and temporal resolutions, associated privacy risks and resource requirements for the collecting device.
GPS sensors report the latitude and longitude of the tracked device with the aid of positioning satellites. While GPS localization is highly accurate (within a few metres), it requires clear line of sight to at least three satellites all the time; its accuracy reduces significantly in regions with obstruction. Moreover, it consumes a lot of battery power and its continuous precise recordings are often privacy invasive. Thus, popular mobile phone ecosystems like Android, iOS have recently put extra precautions for application developers to use this sensor.
Call data record (CDR) records the location of the base station (the cell tower a phone is connected to). It has coarse temporal resolution as the tracking generally takes place when a person uses call, messaging or data service. As a cell tower usually serves a large area its accuracy is of the order of hundreds of metres. Its coarse spatial and temporal resolution makes it less privacy invasive. Wi-Fi access points also can offer similar localization service with increased spatial and temporal resolution. While CDR is suited for outdoor localization for its wide adaptation, Wi-Fi localization is suited for indoor localization for its higher resolution and availability in indoor spaces.
Check-in data record when a person logs visits to venues, for example posts in social media, such as Twitter, Foursquare, Facebook, etc. Check-ins have low temporal granularity and less scaling power to a large population. However, they can capture longitudinal lifestyle with limited data size.
Collocation information is often more important than the location itself, for example, in the study of infectious disease spread. While this can be inferred from the above types of data, it can also be recorded directly by tracking proximity to other devices using Bluetooth or RFID.
2.2.2. Features
Unlike many other sensing modalities (e.g. audio), features from mobility data are not standardized. This is primarily because complex spatio-temporal mobility data are hard to describe in a succinct unique way and different applications need to take varied perspectives.
Simple features characterizing individual trips or the spatial spread of movements are commonly used in the literature (table 2). Techniques to extract features differ based on the type of data at hand; for example, one needs to extract the interesting parts of a GPS trace prior to extracting the features, namely the stay points (where the person stayed beyond a certain duration) and the movement between stay points. Such pre-processing steps are common across applications and often offered as a framework [93].
Table 2.
feature | short description |
---|---|
length of trips | trips are defined specific to applications, for example, movement between two stay points, or using a single mode of transport, etc. |
distance from home | home location is often absent in the data for privacy reasons. It is often inferred using heuristics like the location at night |
different places visited | characterized by the count of unique ones or the frequency distribution represented with entropy |
radius of gyration | deviation from the centroid of the collection of places visited (S): , where C is the centroid location of S |
entropy-based measures | entropy of the distribution of the places visited |
POI-based features | variance in the type of POIs, relative frequencies, etc. |
Health applications typically benefit from interpretable features to build explainable models. Healthwalks [94] builds such a framework that takes user location traces, associated points of interest and user social demographics as input and automatically generates combinations of features using a deterministic automaton (DFA). The effectiveness of the generated features is shown using an eight class disease classification. The disease categories include healthy, mental, dental, neurological, orthopaedic illness, etc.
While the above features are useful in many applications, recently researchers have been using tools from algebraic topology [95] and geometry [96] to map trajectories to points in Euclidean space to make them amenable to ML tools. Further, progress is being made for using trajectories in the ML algorithms [97,98]. However, health applications are yet to take advantage from these progresses.
2.2.3. Applications
Health applications for mobility data range from individual health monitoring and diagnosis to augmenting public health decisions. In the following, we discuss different health applications that use mobility information. We divide the applications into three categories: (i) where mobility affects the disease progression in a population, (ii) where the disease changes mobility patterns in individuals that lead to digital biomarkers for diagnosis and (iii) where monitoring mobility helps in planning and managing health related resource provisioning.
The analysis pipeline differs across applications. The typical pipeline shown in figure 2 applies to the applications in the second and third category. Applications in the first category mostly use collocation data and generally use the data directly without extracting features.
Disease spreading by mobility. Infectious diseases such as COVID-19 and H1N1 are spread primarily by human contacts. The effect of mobility for transmitting such diseases has long been studied, for example in Spanish flu [99]. Due to COVID-19 the study of human mobility and its connection to spreading infectious diseases has surged naturally.
Mobility-based disease spread modelling. While traditional disease spread models like SEIR (susceptible exposed infected recovered) assume contacts between random pairs of individuals, incorporating mobility in the model increases its predictive power. For example, Schwabe et al. [100] proposed mobility marked Hawkes model combining Hawkes process to model the transmission dynamics of the disease and a regularized Poisson regression to incorporate mobility covariates extracted from telecommunication data. The model outperforms the other models in forecasting the infections in the future for 5–21 days in Switzerland.
Mobility interventions. According to the scale of movement, mobility has different effects on disease spreading, for example, fine-grained mobility within a local geographical region in terms of venues visited and number of face-to-face meetups affect the growth rate and demographic patterns of infections in that region. On the contrary, movement between countries or large geographical regions may spread a new disease or a variant to a different country or region. While both are important to study in regard to infectious diseases, fine-grained mobility is popularly captured and studied using mobile devices and is discussed in the article.
As infectious diseases spread via mobility, the most popular non-pharmaceutical interventions to reduce their spread have naturally been lockdowns. Several studies across different countries have shown their effectiveness [101,102]. Using disease simulation on aggregated inter-province movement counts in China as derived from Baidu location based services [101] has shown that the travel restrictions in Wuhan introduced on 23 January 2020 had reduced the number of cases exported internationally by 70–80% until mid February 2020. CDR data are also useful to draw similar conclusions—analysing the data for the most affected counties in the USA [102] found that the decreased mobility strongly correlates (Pearson correlation coefficients above 0.7) with decreased COVID-19 case growth rates.
Although mobility restrictions are effective in flattening the curve they immediately affected the economy and societal functions [103]. While designing the optimal lockdown policy has attracted theoretical interest [104], a more practical approach is to provide a tool for evaluating the possible effects of different intervention strategies [105]. People and venues have non-uniform effect in spreading a disease, e.g. large gatherings act as super-spreaders. Strategies to reopen an economy from lockdowns has been studied [106,107] using check-in data, Wi-Fi association logs and GPS traces. The primary conclusions are that to reduce the disease spread and retain social activities, reopening strategies need to consider the mobility of the people, for example, protect the most mobile people or restricting large gatherings.
Mobility and other diseases. While in the above we reviewed COVID-19 related studies for their timely relevance, mobility restrictions had been used and studied in the past, for example to stop the spread of Ebola in Sierra Leone [108], H1N1 influenza [109], etc. Also contact tracing in the context of disease spreading has long been studied [110]. Apart from the above infectious diseases, mobility has been shown to play a major role in spreading other diseases like malaria [111]—people travelling to different cities carry the parasite and infect the mosquitoes there and in turn the population.
Mobility to aid diagnosis. Several health problems correlate with mobility behaviour changes. Such behavioural changes are useful in predicting or diagnosing a health problem passively. This method of diagnosis is inherently less costly to deploy at large scale than traditional survey and clinical test based methods simply due to pervasive presence of mobile phones and wearable devices in the population.
Mobility as a potential digital biomarker. Spatial navigation patterns are affected by dementia and are considered as one of the earliest symptoms [112]. Pervasive localization sensing in personal devices and progress in related AI techniques have enabled the recent trend in research to use everyday movement data to characterize dementia. One study [113] tracked seven patients and eight healthy people for eight weeks using GPS tracking devices. It was found that the destinations for the dementia patients (with 95% accuracy) are more predictable than the control population (with 92% accuracy). Features extracted from driving trajectories are found to have potential to become digital biomarkers for Alzheimer’s disease (AD) [114]. Using the data from 75 without and 64 people with preclinical AD, the authors built a random forest classifier for identifying preclinical AD patients with F1 score as 0.82. The most important driving based features were average jerk, number of night trips and radius of gyration. Further it has been shown that Alzheimer’s patients have significantly different outdoor mobility behaviour than control population even in early stage [115]: patients explore spatially less compared to controls. A patient’s mobility patterns differ when accompanied by carers.
Individual mobility behaviour changes in fever, influenza or mental stress situations [116]. Behavioural changes are reflected in total communication, interactions with respect to time of day (e.g. late night, early morning), diversity and entropy of face-to-face interactions and movement. These are quantified using co-location and communication data tracked using Bluetooth proximity, call logs and Wi-Fi connections from the participants’ mobile phones. Finally the health status of an individual is predicted with a Bayesian network classifier, without having actual health measurements from the subjects. GPS traces also produce similar conclusions [117]—significant correlation exists between mobility trace characteristics and the depressive moods of individuals.
Other predictive applications. Apart from diagnostic inference at snapshots, navigation lifestyle analysis is helpful in studying longitudinal disease progression [118]. The study exploits the causal relationship between chronic diseases and daily habits captured by check-ins. The points of interest (POIs) are first embedded into 100 dimensional vectors according to their transition patterns capturing their semantic meanings. Categories of the POIs are then clustered using a Gaussian mixture model in a predefined number of clusters. The main insight here is that in a city region the distribution of POI categories is a good predictor of the distribution of progression of chronic diseases. Authors have used collaborative topic modelling to find the latent factor to model the interaction between regions and POI categories and used the same latent factorization to model the region–disease progression interaction.
It is possible to predict hospitalization using mobility behaviour [119]. The study recorded CDR data from over a thousand participants for almost two months. The hospitalization data were collected using surveys. The study used radius of gyration, distribution entropy of visited locations, area of green spaces in the neighbourhood and POI densities near the places visited. Interesting findings include that less mobile people with less visit diversity, and who visit fewer sports facilities are more likely to visit hospitals; fast food store significantly affects health of younger people.
Analysing causality between health status and mobility. However, one should be careful while using the correlation between mobility behaviour and health status; for example, older people visit parks more often but have worse health status than younger people. Thus a correlation analysis without considering age will lead to a negative correlation between park visits and health status. Causal relation between mobility and health status has also been analysed [120]. The proposed causal framework is based on propensity score matching. The framework is extended to health prediction models selecting causal features resulting in a robust model.
Health resource provisioning. Access to resources heavily affects health outcome be it the facilities for health being in the neighbourhood or critical machines in a hospital. In many such scenarios the effect can be better understood and managed by studying the locations of the facilities and their users.
Access to outdoor facilities. Poor physical access to health facilities causes reduced uptake of preventive health services and is likely to affect low-income households the most. A study [121] used cell tower location data from 14 million subscribers tracked for a year in rural Kenya to find the correlation between the mobility of the people, access to healthcare and uptake of two preventive healthcare interventions: childhood immunizations and antenatal care. They characterized the mobility of a person with a single measure, radius of gyration that captures both the frequencies and distances that one travels. They found that long travel times to health facilities strongly correlate with increased mobility in geographically isolated areas. Furthermore, they have shown that the mobile phone-derived measures of mobility can predict which regions are lacking preventive care.
An app was developed [122] for elderly people to locate an accessible and adequate healthcare centre that can attend to their needs. The app uses phone GPS or the cell tower location to track the user location and searches for the nearest facility from a database.
The relation between body weight and lifestyle is intuitive and has been researched using surveys and census data. Burgoine et al. [123] tracked school children for a week with GPS devices and found that greater school walkability was associated with significantly lower mean BMI. Greater home walkability was associated with increased BMI, as was greater school access to green space.
Access to indoor facilities. Indoor positioning is being used in hospitals to track medical devices and staffs for better resource management, enhancing patient experience, and repair and maintenance management [124,125]. Further indoor positioning is used to protect visually impaired, elderly people from collisions [126]. It uses RFID to identify and track passive RFID tags by analysing the received backscatter signals. By extracting the RSSI and phase profiles as fingerprints it estimates the distance to obstacles without observations of eyes and thus guiding the users to move in the domain safely.
2.3. Motion for physical activity and fitness
Measuring physical activity at scale. Large scale studies of physical activity and well-being leveraging mobile devices’ built-in accelerometers have shown promise as global physical activity surveillance tools, demonstrating, for example, inequalities across different countries after analysing data from over 700 000 people [2]. Another analysis of exercise patterns in a global social network of 1.1 million runners demonstrated that exercise is ‘contagious’, whose effect depends on gender and relative levels of activity [127]. Ten million users of a weight monitoring app were used to show that people are more likely to lose weight when they had more friends of the opposite sex [128]. Weight loss was the subject of other studies of over 1 million participants [129], which showed that power users demonstrated the greatest weight loss. The relationship between physical activity and cardiovascular disease (CVD) was studied for 50 000 people in [130], finding that lower overall activity but more frequent transitions between active and inactive periods was associated with similar CVD to higher overall activity but with fewer transitions. Mobile and wearable sensors allow for continuous and ubiquitous monitoring of an individual’s physical activity profiles, which, when combined with cardio-respiratory information, provide valuable insights into that individual’s health and fitness status [131]. Hence, the possibility of measuring individuals’ physiological characteristics in free-living conditions is of great interest for research, clinical and commercial applications. In particular, physical activity is characterized by both movement and the associated cardiovascular response to movement (e.g. heart rate increases after exercise and the dynamics of this increase are dictated by fitness levels [132]); thus, leveraging these two signals concurrently likely produces better representations than either signal taken in isolation. For instance, heart rate (HR) responses to exercise have been shown to be strongly predictive of CVD, coronary heart disease (CHD) and all-cause mortality [133]. In healthy individuals, HR responses to activity are defined by an increase in HR that is concurrent to the increasing intensity of the activity [134]. More recently, the Apple Watch has introduced cardio-respiratory fitness (CRF) features [135] while large population-scale studies showed that these HR responses can predict VO2max [136].
2.3.1. Data
Challenges in modelling wearable data. The advent of wearable technologies has given individuals the opportunity to unobtrusively track everyday behaviour. Given the rapid growth in adoption of Internet-enabled wearable devices, sensor time-series comprise a considerable amount of user-generated data [137]. However, extracting meaning from these data can be challenging, since sensors measure low-level signals (e.g. acceleration) as opposed to the more high-level events that are usually of interest (e.g. arrhythmia, infection or obesity onset). Most wearable devices, particularly those that are wrist-worn, incorporate accelerometry sensors, which are very affordable tools to objectively study physical activity patterns [17]. However, since wearables are used in daily, unconstrained environments, activities such as drinking coffee or alcohol, as well as stress, may confound simple heuristics.
2.3.2. Modeling
Modelling wearable signals with ML. ML models have been only recently applied to this task. To approximate VO2max without the need for a dynamic test, non-exercise models aim to provide a viable alternative to CRF assessment for widespread use in many healthcare settings. These are usually traditional regression models and incorporate variables like sex, age, body mass index (BMI), RHR and self-reported physical activity to infer VO2max [138]. However, the validity of such estimation is still much lower than what can be achieved with dynamic exercise testing [139]. Wearable devices, such as activity trackers and smartwatches, increasingly provide opportunities for non-intrusive objective monitoring of biological signals such as HR and movement during free-living, potentially enabling more precise prediction of VO2max without the need to conduct a specific exercise test [140]. A recent large-scale analysis used data from Fenland Study (N = 11 059), along with its longitudinal cohort (N = 2675), and a third external cohort using the UK Biobank Validation Study (N = 181) who underwent maximal VO2max testing, to show that the combination of wearables and other biomarkers as inputs to neural networks yields a strong correlation to ground truth in a holdout sample (r = 0.82, ) [136]. It also detects fitness change over time (e.g. after 7 years). Notably, the model’s latent space can also be used for fitness-aware patient subtyping paving the way to scalable interventions and personalized trial recruitment. These results demonstrate the value of wearables for fitness estimation that today can be measured only with laboratory tests.
Recent advances in deep learning architectures for sequential modelling based upon wearable and mobile sensing have been used for health predictions and recommendations [141]. For example, FitRec, an LSTM-based approach to modelling HR and activity data for personalized fitness recommendations, was able to learn activity-specific contextual and personalized dynamics of individual user HR profiles during exercise segments [56]. This approach is helpful but requires prior segmentation of activities, which can be a constraint when applying these techniques in free-living, unconstrained conditions. Recent work using self-supervised learning has shown promise in the same data modalities such as ECG data [142,143]. Further, we recently developed a self-supervised model which exploits the multi-modal data of modern wearables to learn meaningful representations which generalize to several outcomes with transfer learning [144]. The model maps activity data to future HR responses and can be used as a feature extractor for wearable data. The value captured by the learned embeddings was showcased through strong performance at inferring physiologically meaningful variables ranging from anthropometrics to fitness, outperforming unimodal autoencoders and common biomarkers. For instance, the embeddings achieved an AUC of 0.68 for cardio-fitness prediction and an AUC of 0.80 for physical activity energy expenditure. Obtaining these outcomes in large populations can be valuable for downstream health-related inferences that would normally be costly and burdensome.
The emergence of large foundation AI models including large language models (LLMs) holds a promise to revolutionize this area. Self-supervised learning uses existing data to learn meaningful patterns by artificially creating predictive tasks such as predicting the next value or filling the gaps. In the context of health signals, these tasks can range from predicting the arrow of time, to contrastive learning and masking [145,146]. Masking, in particular, has emerged as a universal prediction task in LLMs where widely successful models such as ChatGPT are trained to predict the next word. Very recent work has demonstrated that by prompting LLMs with wearable time-series data and applying few-shot tuning, they are capable of making meaningful inferences on numerous health tasks for both clinical and wellness contexts [147].
2.3.3. Applications
Fitness as a key application. A concept that is particularly central to physical activity is that of CRF, an important modifiable marker of cardiovascular health embodied by a strong inverse relationship with the incidence of CVD, type 2 diabetes, cancer, mortality and other adverse health outcomes [131]. Clinical evidence shows that CRF is not only potentially a stronger predictor of mortality than well-established risk factors like hypertension, type 2 diabetes, high cholesterol or smoking, but that using CRF to complement these traditional risk factors significantly improves the precision of risk prediction models for adverse CVD health outcomes [148]. Beyond its implications in medicine, CRF is frequently used in sports as an indicator of endurance capacity, having strong predictive value for other sport-related performance traits [148].
Challenges in collecting fitness ground truth. The gold-standard measure of CRF is the maximal oxygen uptake (VO2max), which measures the maximal rate at which an individual can consume oxygen during exercise. VO2max is assessed through an exercise test to exhaustion while respiratory gas exchange is measured, with the assessment only deemed a true maximal result if several test criteria are met. These criteria include levelling-off of oxygen uptake and HR and the surpassing of thresholds for the respiratory exchange ratio. This type of assessment requires trained staff and expensive laboratory settings with specialized equipment and often test criteria for maximal effort are not met [149]. Given these logistical constraints and the inherent risk of maximal exercise testing, scalability of fitness assessment in large populations has been limited, meaning relatively little is known about population levels of fitness, or their possible changes over time. Despite some promising studies which attempt to infer VO2max from data collected during free-living conditions, these mostly stem from small-scale cohorts with less than 50 participants and use contextual data from treadmill activity, which again limits their application in real-world contexts [150].
2.4. Motion for mental health
2.4.1. Data
Data and outcomes. Mood and general mental wellbeing have been associated with several clinical outcomes. Self-reported sadness was found to be an indicator of depression [151], while self-reported happiness is linked to longevity [152], personality traits [153] and reduced mortality risk [154]. The experience sampling method (ESM) or ecological momentary assessment (EMA)—which involves asking participants to report their behaviours or environment on repeated occasions over time—has long been used as a mechanism to longitudinally assess the mental health of individuals by prompting them to report their mental state using questionnaires, traditionally administered using pen and paper and also via the Web. Psychologists have used different tools or scales to measure mood. These include the positive and negative affect schedule (PANAS) [155], a self-report questionnaire of two 10-item scales that measures both positive and negative affect; and the Affect Grid scale [156], a two-dimensional grid, where the x-axis indicates the feeling in terms of its positiveness or negativeness and the y-axis indicates its intensity. Independently of the scale used, timely and accurate mood reporting is important to anticipate clinical outcomes. To this end, smartphones and wearable devices have enabled timely delivery of experience sampling [157], allowing a near real-time detection of clinical outcomes and relapses.
2.4.2. Modeling
ML for mental health. Regarding ML models, the majority of related literature has applied some kind of supervised learning, such as logistic regression or SVMs, which cannot capture nonlinear combinations of features. For a more extensive view of the field, we point the interested reader to a comprehensive review on ML for mental health [158]. Some recent works employ recurrent neural networks [20], feed-forward layers [159], multi-task learning [160] and autoencoders to fill in missing sensor data [161], or to learn better representations [162]. Further, binary prediction is quite common in the mood prediction literature, where mood is simplified to a binary state [21], so that, for instance, extreme depression is binned in the same class as moderate unhappiness. Since neutral mood might be uninformative and make the predictions harder, authors often omit the middle-scoring 40–60% of reports. Instead, recent work explored fine-grained mood prediction through regression and clustering [163], as well as novel multi-dimensional formulations of the Affect Grid through multi-task learning [164].
2.4.3. Applications
Multiple applications of motion data for mental health have been proposed from both research and commercial ventures. For instance, StudentLife [63] combined sensing and self-reports to assess the impact of student workload on stress, whereas Snapshot [160] tracked their mood and sleep. Others focused on detecting depression by tracking medication, sleep patterns and actions [20], location [11] or keypress acceleration [165]. On a larger scale, Utsureko [20] and EmotionSense [21], two independent smartphone applications for mood monitoring through self-reports, were used by more than 24 000 and 17 000 users, respectively. On the other hand, Evidation Health, a private company, offers users money for contributing their sensor data along with self-reported surveys.
2.5. Other modalities
Other widely adopted sensing modalities for mobile health can be grouped into two main streams: (i) biosignal sensors and (ii) radio frequency (RF) sensing. The focus of this paper is on movement, location and audio, but we also want to acknowledge that there are other data modalities and sources that can be useful for health sensing.
Biosignals. The two most widely adopted sensor data types are ECG and photoplethysmography (PPG) [166–168]. ECG sensors are generally placed on the chest to measure the electrical activity generated by heart muscle contractions directly. They can collect reliable cardiac information for heart monitoring, e.g. abnormal heart rhythms, and detection of atrial fibrillation. PPG sensors are an alternative non-invasive way to monitor heart rates, which shine light on the skin surface and measure the changes in blood circulation. This information is valuable and can be used to infer the abnormality in heart rates. Due to their non-invasive and easy-to-carry nature, PPG sensors have been widely employed on smartwatches, such as the Apple Watch, to monitor cardiovascular health [169]. However, they may still have their own limitations, such as being bulky to use in daily life (e.g. ECG sensors in a chest strap), inaccurate measurement under motions (e.g. PPG sensors), sensitivity to environmental conditions, high power consumption, etc. [170]. Apart from these two types of sensors, there are also other more complex sensing modalities such as electroencephalography (EEG) and electrooculography (EOG) [171], e.g. for sleep monitoring, but they are less employed in consumer wearable and mobile devices, rather in only clinical devices that limit their employment.
RF signals. Bluetooth and Wi-Fi signals on mobile devices are two of the main RF signal types for mobile health. Many mobile apps have been developed using such technology to understand outbreak epidemiology, individual screening and contact tracing [172–174]. Many of these apps use Bluetooth and Bluetooth Low Energy to infer proximity between people. Multiple governments worldwide and industries have released contact tracing apps for mobile phones. For example, Singapore implemented a mobile phone application, TraceTogether, which pushes notifications to devices in close proximity as 2 m apart and sharing information [175]. Some similar apps in different countries have also been employed for COVID-19 tracing [176,177].
However, due to false negatives in detecting face-to-face contacts, digital contact tracing may not be as effective as its manual counterpart—a study [178] estimated that digital tracing alone can reduce the R number (average number of people infected by a single infectious person) by whereas manual tracing of all contacts can reduce R number by . Further, the apps lack in transparent quantification of individual advantage in obeying the app’s advice [179]. Moreover, people without economic privilege have a limited capacity to self-isolate at home, which undermines the effectiveness of these approaches.
Both Bluetooth and Wi-Fi signals have been widely used for indoor localization particularly for older age groups, in assisting their living quality and also the emergency response [180,181]. Wi-Fi signal is one of the key modalities for human activity recognition and gesture recognition [182]. As Wi-Fi can travel through almost any indoor corner, human body movement or gesture changes can lead to small changes in signal propagation, and thus can be used to detect activities. Recently, RF signal has also been extended for sleep posture monitoring [183], which transmits the radio signals and observes their reflections from the surroundings. It provides non-contact sleep monitoring solutions.
3. Implications and limitations
3.1. Limitation for current state of AI
Interpretability. A significant hurdle that hinders the wide-scale adoption of AI systems in healthcare is the lack of interpretability, making it challenging for users and medical practitioners to trust these systems. Despite previous attempts to address the challenges surrounding model interpretability, most solutions have been implemented post hoc (such as the attention mechanism) focusing on explanations in certain domains, such as vision [184]. For instance, while sequential deep learning can generate fairly accurate predictions when dealing with motion sequences or audio features in longitudinal time series, it remains difficult to interpret how the model makes such judgments based on historical samples. Another potential solution for enhancing interpretability is adopting uncertainty-aware modelling approaches. These enable the prediction of uncertainty associated with the final outcome predictions, thereby providing confidence levels for users to either trust or distrust the systems. One study [185] employed ensemble learning for sound-based COVID-19 detection to infer the uncertainty of the model predictions and showed that false predictions often yield higher uncertainty, enabling us to suggest to users with certainty higher than a threshold to repeat the audio test on their phones or to take clinical tests if digital diagnosis still fails. Furthermore, active learning methods could enable the selection of uncertain samples or users for further analysis such as an additional review by domain experts (physicians) [186]. Evidence also suggests that while deep learning can identify spurious correlations, it often fails to determine causation, which can be essential in a clinical context [187]. Medical practitioners need to understand the causal factors influencing outcomes rather than merely observing correlations. There is an emerging trend away from solely favouring straightforward ML systems like linear regression or SVMs to ensure explainability or overlooking interpretability in favour of performance gains using opaque deep learning models. New opportunities are blossoming in the realm of interpretability, not only encompassing previously discussed areas but also highlighting the potential of integrating domain knowledge from clinicians into modelling approaches [188].
Adaptation and personalization. Data are persistently accumulated over time while monitoring an individual’s health. Owing to factors like ageing or disease progression (e.g. deterioration), the attributes of such data can shift, a phenomenon referred to as the data distribution shift. Models trained on data gathered from a specific temporal cutoff may underperform in later data collection. Consequently, updating these models with new data becomes vital. Additionally, different individuals may exhibit varying symptoms or responses to the same treatment. This requires meticulous attention for systems designed around individual needs, making a case for improved personalization [189]. Such systems can adapt and cater to the heterogeneity among users, further enhancing the effectiveness of disease monitoring and healthcare interventions. Although existing methodologies like continual learning or transfer learning may present potential avenues for adaptation and personalization such as for electronic health record [190], their exploration and application in the context of mobile health remain relatively underdeveloped.
Fairness and robustness. Fairness ensures that algorithmic decisions do not create discriminatory or unjust results when comparing different demographics (e.g. race or sex). Real-world cases of ‘unfair’ ML algorithms abound, such as skin lesion classification models underperforming when assessing black patients, compared to white patients. At the same time, people of colour are consistently misclassified by health sensors such as oximeters as they were validated on predominantly white populations [191]. Applying ideas from ML fairness could possibly solve some of these problems; however, sensing data and models have certain particularities. For example, they typically deal with small-scale studies, proof-of-concept datasets often collected in the laboratory or in the wild. Contrary to the tabular format of such datasets, sensing data are mostly sequential in nature, with biases being harder to surface. In other words, while it is relatively straightforward to distinguish a person’s skin tone given a picture, it is much harder to do so from oximetry measurements, necessitating the collection of supplementary metadata, such as demographics. Specifically, we should consider groups relevant to a particular scenario or deployment setting. For example, in a depression screening application, gender could be relevant as a sensitive attribute due to reported gender differences in the disorder’s signals [192]. Similarly, data validation methods can facilitate fairness and robustness and, therefore, should also be applied across sensitive attributes. For example, visualization tools such as the What-If Tool [193] and FairLens [194] can help surface any potential data outliers and help correct them before they creep into models. More technically, the literature is increasingly interested in approaches that mitigate biases. Considering that most outcomes in mobile health sensing are numerical (regression tasks), Agarwal et al. [195] introduced two definitions of fairness in regression: statistical parity, which requires the prediction to be statistically independent of the sensitive attribute, and bounded group loss, in which the prediction error is restricted to any sensitive group to remain below some predefined threshold. Despite recent efforts within the mobile and ubiquitous computing community [196], fairness and robustness are still underexplored with regard to mobile health sensing. Robustness to distribution shifts can also be dealt with using domain adaptation or personalization techniques that are tailored to individuals and learn from their unique preferences (see previous paragraph).
3.2. Limitation for mobiles and wearables
Resource constraint. As larger models like GPT4 and other LLMs become more prevalent due to their improved capabilities, we can expect specialized foundation models tailored to healthcare applications [197]. These applications are increasingly relying on domain-specific models at the moment. This task-agnostic shift, however, presents a mounting challenge in deploying these systems on resource-limited devices such as smartphones and wearable devices, especially when considering aspects like memory capacity and battery life. Some efforts have been made on both the software and hardware to optimize deep learning inference, from model pruning and compression techniques that reduce model size and complexity to parallel network computations [198]. However, when it requires fine-tuning or retraining the models on edge devices, it becomes a challenge. This limits its applications in various contexts, e.g. personalizing the model with more new user data input, or making continual adaptation with new data. Thus, the focus is shifting towards efficient deep learning training on edge devices. Instead of transmitting local device data to a server (a process that is bandwidth-intensive and raises privacy issues), the emphasis is on distributed training methods. Typically, there is one central server paired with multiple edge devices. The breakthroughs in efficient deep learning training revolve around optimizing training update frequencies between local and server devices, determining which part of the model needs an update, and so forth [199]. Despite progress and even with advanced transformers [200], hurdles related to latency, energy consumption and compute resource management persist. Although the existing studies have not been tailored for mobile health, they share similar challenges. Hence, determining how to train and deploy deep learning models with the full potential of such resource-limited wearable devices merits thoughtful exploration.
Data and label quality. Very few phones and wearable devices have medical-grade sensors and are costly. The majority of devices on the market have a cheap sensor which produces noisy data or is easily interfered with by motion artefacts. This requires specific attention in data processing or model design. Proper denoising techniques before passing data for modelling or specific model structures such as multi-task learning have been validated to denoise the signals or eliminate motion artefacts. Moreover, sensor data might be missing during motions, or data are sampled at irregular time intervals in a longitudinal way. These characteristics may require additional mechanisms in AI algorithms to handle effectively. On the other hand, label quality should also be carefully examined before modelling. To make the study scalable, some labels are collected in a crowdsourced way, by self-reporting. Therefore, either further automation in the label quality check or manual check is required.
3.3. Limited validation of the methodology
Limited datasets. Data collection is time-consuming and cost-intensive. Most mobile health apps are validated only using a limited number of participants. For example, Apple’s CRF feature is publicly validated on 221 participants [135] while non-commercial methods use population-scale data from more than 11 000 participants [136]. Considering the large user bases of such products, we need to find better methods to leverage free-living with small-scale laboratory-collected data.
Lack of validation standardization for deployment. New drugs or treatments undergo strict certification and regulatory screenings before they become available to the public. Most commonly, the vendors conduct randomized controlled trials in order to rule out side effects and validate the efficacy thereof. In mobile health, validation is still a thorny issue. Data remain siloed within organizations and companies and most of the time treated as a ‘moat’, or a competitive advantage of each company. The few mobile health products that become available are labelled as ‘well-being’ apps so as to bypass the strict regulatory screening for medical devices. Mobile health sensing is an emerging area that should adopt more rigorous evaluation protocols from the medical AI community [201].
3.4. Privacy and security
Last, sensing signals from mobile phones often contain privacy-sensitive information. For example, continuously recorded audio from phones will reveal personal conversations, and mobility data can leak sensitive location visits and meetups. However, not every sensing modality poses the same threat; for example, motion data are comparatively less privacy sensitive. On the other hand, mobility data expose sensitive personal information. Moreover, security has always been a significant challenge in mobile health applications. Several studies have identified security challenges that affect the development of secure mobile health applications. For example, out of 20 000 apps from Google Play studied in [202] it was revealed that 45% of the apps rely on unencrypted communication, and as much as of personal data (e.g. location information and passwords) are sent on unsecured traffic.
Both privacy and security are popular topics of research and several approaches have been proposed including anonymity frameworks [203] to more recently proposed differential privacy [204], and blockchain-based solutions [205]. More recently ML on decentralized data in terms of federated learning has shown promise to preserve privacy while building useful models out of the data [206,207]. Another popular way to mitigate privacy risks is to use on-device computation. For example, consider a location-tracking diagnostic application. A potential model using datasets acquired from multiple individuals with consent. Then the classifier can be deployed in an individual’s phone to run the inference on the user’s sensitive location data without sharing it with a third party.
While security and privacy in mobile health applications are of particular importance, we point the reader to other material that is available [208,209].
4. Conclusion
This paper has thoroughly examined the possibilities and challenges associated with human-centred AI for mobile health sensing, specifically focusing on three primary sensing modalities: audio, mobility and motion data. It has provided a comprehensive coverage of various aspects, ranging from data collection methods to modelling pipelines, and explored potential applications in depth. Furthermore, it has critically addressed the limitations of current approaches and proposed potential solutions, encompassing wearable devices, AI modelling and validation, while shedding light on privacy and security concerns.
The objective of this paper is to present a holistic perspective and offer valuable insights into the realm of mobile health sensing. Its aim is to enhance the understanding of human-centred AI and inspire advancements in interdisciplinary fields for future progress.
Footnotes
In this paper, we use the terms ML and AI interchangeably, with ML describing the development of algorithms and statistical models such as feature engineering or unsupervised learning, and AI describing the systems or machines in general that perform tasks requiring human intelligence.
We define mobile sensing as data from connected sensors which are used to characterize behaviours related to health. Other terms used in the literature are personal sensing, digital phenotyping and context sensing, with different albeit overlapping connotations. We point the interested reader to the discussion in [12].
Ethics
This work did not require ethical approval from a human subject or animal welfare committee.
Data accessibility
This article has no additional data.
Declaration of AI use
We have not used AI-assisted technologies in creating this article.
Authors' contributions
T.D.: conceptualization, writing—original draft; D.S.: conceptualization, writing—original draft; A.G.: conceptualization, writing—original draft; C.M.: conceptualization, funding acquisition, supervision, writing—review and editing.
All authors gave final approval for publication and agreed to be held accountable for the work performed therein.
Conflict of interest declaration
We declare we have no competing interests.
Funding
This work was supported by Wellcome Trust Project 213939 and ERC Project 833296.
References
- 1.Kim ES, Park N, Sun JK, Smith J, Peterson C. 2014. Life satisfaction and frequency of doctor visits. Psychosom. Med. 76, 86. ( 10.1097/PSY.0000000000000024) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Althoff T, Sosic R, Hicks JL, King AC, Delp SL, Leskovec J. 2017. Large-scale physical activity data reveal worldwide activity inequality. Nature 547, 336-339. ( 10.1038/nature23018) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Tejani N, Dresselhaus TR, Weinger MB. 2010. Development of a hand-held computer platform for real-time behavioral assessment of physicians and nurses. J. Biomed. Inform. 43, 75-80. ( 10.1016/j.jbi.2009.08.011) [DOI] [PubMed] [Google Scholar]
- 4.Van Berkel N, Ferreira D, Kostakos V. 2018. The experience sampling method on mobile devices. ACM Comput. Surv. 50, 93. ( 10.1145/3123988) [DOI] [Google Scholar]
- 5.He K, Zhang X, Ren S, Sun J. 2016. Deep residual learning for image recognition. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition, pp. 770–778.
- 6.Schrittwieser J, et al. 2020. Mastering atari, go, chess and shogi by planning with a learned model. Nature 588, 604-609. ( 10.1038/s41586-020-03051-4) [DOI] [PubMed] [Google Scholar]
- 7.McKinney SM, et al. 2020. International evaluation of an AI system for breast cancer screening. Nature 577, 89-94. ( 10.1038/s41586-019-1799-6) [DOI] [PubMed] [Google Scholar]
- 8.World Health Organization. 2002. The world health report 2002: reducing risks, promoting healthy life. Geneva, Switzerland: World Health Organization. [Google Scholar]
- 9.Strain T, Wijndaele K, Dempsey PC, Sharp SJ, Pearce M, Jeon J, Lindsay T, Wareham N, Brage S. 2020. Wearable-device-measured physical activity and future health risk. Nat. Med. 26, 1385-1391. ( 10.1038/s41591-020-1012-3) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Han J, et al. 2021. Sounds of COVID-19: exploring realistic performance of audio-based digital testing. npj Digit. Med. 5, 16. ( 10.1038/s41746-021-00553-x) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Palmius N, Tsanas A, Saunders KE, Bilderbeck AC, Geddes JR, Goodwin GM, De Vos M. 2016. Detecting bipolar depression from geographic location data. IEEE Trans. Biomed. Eng. 64, 1761-1771. ( 10.1109/TBME.2016.2611862) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Mohr DC, Shilton K, Hotopf M. 2020. Digital phenotyping, behavioral sensing, or personal sensing: names and transparency in the digital age. npj Digit. Med. 3, 45. ( 10.1038/s41746-020-0251-5) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Ang WT, Khosla PK, Riviere CN. 2007. Nonlinear regression model of a low-g MEMS accelerometer. IEEE Sensors J. 7, 81-88. ( 10.1109/JSEN.2006.886995) [DOI] [Google Scholar]
- 14.Lane ND, Miluzzo E, Lu H, Peebles D, Choudhury T, Campbell AT. 2010. A survey of mobile phone sensing. IEEE Commun. Mag. 48, 140-150. ( 10.1109/MCOM.2010.5560598) [DOI] [Google Scholar]
- 15.Radu V, Tong C, Bhattacharya S, Lane ND, Mascolo C, Marina MK, Kawsar F. 2018. Multimodal deep learning for activity and context recognition. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 1, 1-27. ( 10.1145/3161174) [DOI] [Google Scholar]
- 16.Brenner PS, DeLamater JD. 2014. Social desirability bias in self-reports of physical activity: is an exercise identity the culprit? Soc. Indic. Res. 117, 489-504. ( 10.1007/s11205-013-0359-y) [DOI] [Google Scholar]
- 17.Doherty A, et al. 2017. Large scale population assessment of physical activity using wrist worn accelerometers: the UK Biobank study. PLoS ONE 12, e0169649. ( 10.1371/journal.pone.0169649) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Perez MV, et al. 2019. Large-scale assessment of a smartwatch to identify atrial fibrillation. N. Engl. J. Med. 381, 1909-1917. ( 10.1056/NEJMoa1901183) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.O’Connor L, Brage S, Griffin SJ, Wareham NJ, Forouhi NG. 2015. The cross-sectional association between snacking behaviour and measures of adiposity: the Fenland Study, UK. Br. J. Nutr. 114, 1286-1293. ( 10.1017/S000711451500269X) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Suhara Y, Xu Y, Pentland A. 2017. Deepmood: forecasting depressed mood based on self-reported histories via recurrent neural networks. In Proc. 26th Int. Conf. on World Wide Web, pp. 715–724.
- 21.Servia-Rodríguez S, Rachuri KK, Mascolo C, Rentfrow PJ, Lathia N, Sandstrom GM. 2017. Mobile sensing at the service of mental well-being: a large-scale longitudinal study. In Proc. 26th Int. Conf. on World Wide Web, pp. 103–112. International World Wide Web Conferences Steering Committee.
- 22.Radin JM, Wineinger NE, Topol EJ, Steinhubl SR. 2020. Harnessing wearable device data to improve state-level real-time surveillance of influenza-like illness in the USA: a population-based study. Lancet Digit. Health 2, e85-e93. ( 10.1016/S2589-7500(19)30222-5) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Zhang X, Zhou Y, Ma Y, Chen BC, Zhang L, Agarwal D. 2016. Glmix: generalized linear mixed models for large-scale response prediction. In Proc. 22nd ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining, pp. 363–372. New York, NY: ACM.
- 24.Asadzadeh A, Kalankesh LR. 2021. A scope of mobile health solutions in COVID-19 pandemics. Inform. Med. Unlocked 23, 100558. ( 10.1016/j.imu.2021.100558) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Bo F, Yerebakan M, Dai Y, Wang W, Li J, Hu B, Gao S. 2022. IMU-based monitoring for assistive diagnosis and management of IoHT: a review. Healthcare 10, 1210. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Lai S, Farnham A, Ruktanonchai NW, Tatem AJ. 2019. Measuring mobility, disease connectivity and individual risk: a review of using mobile phone data and mHealth for travel medicine. J. Travel Med. 26, taz019. ( 10.1093/jtm/taz019) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Larson EC, Lee T, Liu S, Rosenfeld M, Patel SN. 2011. Accurate and privacy preserving cough sensing using a low-cost microphone. In Proc. 13th Int. Conf. on Ubiquitous Computing, pp. 375–384.
- 28.Markos C, Yu JJ, Da Xu RY. 2020. Capturing uncertainty in unsupervised GPS trajectory segmentation using Bayesian deep learning. In Proc. AAAI Conf. on Artificial Intelligence, vol. 35, pp. 390–398.
- 29.Coppock H, Gaskell A, Tzirakis P, Baird A, Jones L, Schuller B. 2021. End-to-end convolutional neural network enables COVID-19 detection from breath and cough audio: a pilot study. BMJ Innov. 7, 356-362. ( 10.1136/bmjinnov-2021-000668) [DOI] [PubMed] [Google Scholar]
- 30.Ren Y, Wang C, Yang J, Chen Y. 2015. Fine-grained sleep monitoring: hearing your breathing with smartphones. In 2015 IEEE Conf. on Computer Communications (INFOCOM), pp. 1194–1202. IEEE.
- 31.Jiang Y, Li W, Hossain MS, Chen M, Alelaiwi A, Al-Hammadi M. 2020. A snapshot research and implementation of multimodal information fusion for data-driven emotion recognition. Inf. Fusion 53, 209-221. ( 10.1016/j.inffus.2019.06.019) [DOI] [Google Scholar]
- 32.Kim S, Hori T, Watanabe S. 2017. Joint CTC-attention based end-to-end speech recognition using multi-task learning. In 2017 IEEE Int. Conf. on Acoustics, Speech and Signal Processing (ICASSP), pp. 4835–4839. IEEE.
- 33.Lavrentyeva G, Novoselov S, Malykh E, Kozlov A, Kudashev O, Shchemelinin V. 2017. Audio replay attack detection with deep learning frameworks. In Proc. Interspeech 2017, pp. 82–86. ( 10.21437/Interspeech.2017-360) [DOI]
- 34.Tripathi S, Tripathi S, Beigi H. 2018. Multi-modal emotion recognition on iemocap dataset using deep learning. (http://arxiv.org/abs/1804.05788)
- 35.Trigeorgis G, Ringeval F, Brueckner R, Marchi E, Nicolaou MA, Schuller B, Zafeiriou S. 2016. Adieu features? End-to-end speech emotion recognition using a deep convolutional recurrent network. In 2016 IEEE Int. Conf. on Acoustics, Speech and Signal Processing (ICASSP), pp. 5200–5204. IEEE.
- 36.Tzirakis P, Zhang J, Schuller BW. 2018. End-to-end speech emotion recognition using deep neural networks. In 2018 IEEE Int. Conf. on Acoustics, Speech and Signal Processing (ICASSP), pp. 5089–5093. IEEE.
- 37.Valstar M, Schuller B, Smith K, Eyben F, Jiang B, Bilakhia S, Schnieder S, Cowie R, Pantic M. 2013. Avec 2013: the continuous audio/visual emotion and depression recognition challenge. In Proc. 3rd ACM Int. Workshop on Audio/Visual Emotion Challenge, pp. 3–10.
- 38.Dham S, Sharma A, Dhall A. 2017. Depression scale recognition from audio, visual and text analysis. (http://arxiv.org/abs/1709.05865)
- 39.Ringeval F, et al. 2019. AVEC 2019 workshop and challenge: state-of-mind, detecting depression with AI, and cross-cultural affect recognition. In Proc. 9th Int. Conf. on Audio/Visual Emotion Challenge and Workshop, pp. 3–12.
- 40.Martinc M, Pollak S. 2020. Tackling the ADReSS challenge: a multimodal approach to the automated recognition of Alzheimer’s dementia. In Proc. Interspeech 2020, pp. 2157–2161. ( 10.21437/Interspeech.2020-2202) [DOI]
- 41.Rohanian M, Hough J, Purver M. 2021. Multi-modal fusion with gating using audio, lexical and disfluency features for Alzheimer’s dementia recognition from spontaneous speech. (http://arxiv.org/abs/2106.09668)
- 42.Syed MSS, Syed ZS, Lech M, Pirogova E. 2020. Automated screening for Alzheimer’s dementia through spontaneous speech. In Proc. Interspeech 2020, pp. 2222–2226. ( 10.21437/Interspeech.2020-3158) [DOI]
- 43.Amrulloh Y, Abeyratne U, Swarnkar V, Triasih R. 2015. Cough sound analysis for pneumonia and asthma classification in pediatric population. In 2015 6th Int. Conf. on Intelligent Systems, Modelling and Simulation, pp. 127–131. IEEE. [DOI] [PubMed]
- 44.Chen G, Imtiaz SA, Aguilar-Pelaez E, Rodriguez-Villegas E. 2015. Algorithm for heart rate extraction in a novel wearable acoustic sensor. Healthcare Technol. Lett. 2, 28-33. ( 10.1049/htl.2014.0095) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Kusche R, Klimach P, Malhotra A, Kaufmann S, Ryschka M. 2015. An in-ear pulse wave velocity measurement system using heart sounds as time reference. Curr. Dir. Biomed. Eng. 1, 366-370. ( 10.1515/cdbme-2015-0090) [DOI] [Google Scholar]
- 46.Pressler GA, Mansfield JP, Pasterkamp H, Wodicka GR. 2004. Detection of respiratory sounds at the external ear. IEEE Trans. Biomed. Eng. 51, 2089-2096. ( 10.1109/TBME.2004.836525) [DOI] [PubMed] [Google Scholar]
- 47.Amft O, Stäger M, Lukowicz P, Tröster G. 2005. Analysis of chewing sounds for dietary monitoring. In UbiComp 2005: ubiquitous computing (eds Beigl M, Intille S, Rekimoto J, Tokuda H), pp. 56-72. Berlin, Germany: Springer. ( 10.1007/11551201_4) [DOI] [Google Scholar]
- 48.Sharma N, Krishnan P, Kumar R, Ramoji S, Chetupalli SR, Ghosh PK, Ganapathy S. 2020. Coswara—a database of breathing, cough, and voice sounds for COVID-19 diagnosis. (http://arxiv.org/abs/2005.10548)
- 49.Xia T, et al. 2021. COVID-19 sounds: a large-scale audio dataset for digital respiratory screening. In 35th Conf. on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2).
- 50.Ringeval F, Sonderegger A, Sauer J, Lalanne D. 2013. Introducing the RECOLA multimodal corpus of remote collaborative and affective interactions. In 2013 10th IEEE Int. Conf. and Workshops on Automatic Face and Gesture Recognition (FG), pp. 1–8. IEEE.
- 51.Lotfi R, Tzanetakis G, Eskicioglu R, Irani P. 2020. A comparison between audio and IMU data to detect chewing events based on an earable device. In Proc. 11th Augmented Human Int. Conf., pp. 1–8.
- 52.COVID-19 Community Mobility Reports. https://www.google.com/covid19/mobility/index.html?hl=en (accessed 18 October 2023).
- 53.SafeGraph. https://www.safegraph.com/blog/safegraph-partners-with-dewey (accessed 18 October 2023).
- 54.Xu X, et al. 2022. GLOBEM dataset: multi-year datasets for longitudinal human behavior modeling generalization. Adv. Neural Inf. Process. Syst. 35, 24 655-24 692. [Google Scholar]
- 55.Chen X, Wang R, Zee P, Lutsey PL, Javaheri S, Alcántara C, Jackson CL, Williams MA, Redline S. 2015. Racial/ethnic differences in sleep disturbances: the Multi-Ethnic Study of Atherosclerosis (MESA). Sleep 38, 877-888. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Ni J, Muhlstein L, McAuley J. 2019. Modeling heart rate and activity data for personalized fitness recommendation. In The World Wide Web Conf., pp. 1343–1353.
- 57.Levelt WJ, Roelofs A, Meyer AS. 1999. A theory of lexical access in speech production. Behav. Brain Sci. 22, 61-69. ( 10.1017/S0140525X99451775) [DOI] [PubMed] [Google Scholar]
- 58.Davidson RJ, Pizzagalli D, Nitschke JB, Putnam K. 2002. Depression: perspectives from affective neuroscience. Annu. Rev. Psychol. 53, 545-574. ( 10.1146/annurev.psych.53.100901.135148) [DOI] [PubMed] [Google Scholar]
- 59.France DJ, Shiavi RG, Silverman S, Silverman M, Wilkes M. 2000. Acoustical properties of speech as indicators of depression and suicidal risk. IEEE Trans. Biomed. Eng. 47, 829-837. ( 10.1109/10.846676) [DOI] [PubMed] [Google Scholar]
- 60.Cummins N, Scherer S, Krajewski J, Schnieder S, Epps J, Quatieri TF. 2015. A review of depression and suicide risk assessment using speech analysis. Speech Commun. 71, 10-49. ( 10.1016/j.specom.2015.03.004) [DOI] [Google Scholar]
- 61.Tracey B, et al. 2021. Voice biomarkers of recovery from acute respiratory illness. IEEE J. Biomed. Health Inform. 26, 2787-2795. ( 10.1109/JBHI.2021.3137050) [DOI] [PubMed] [Google Scholar]
- 62.Dubey H, Goldberg JC, Abtahi M, Mahler L, Mankodiya K. 2015. EchoWear: smartwatch technology for voice and speech treatments of patients with Parkinson’s disease. In Proc. Conf. on Wireless Health, Bethesda, MD, USA, 14–16 October 2015, article 15. ( 10.1145/2811780.2811957) [DOI]
- 63.Wang R, Chen F, Chen Z, Li T, Harari G, Tignor S, Zhou X, Ben-Zeev D, Campbell AT. 2014. StudentLife: assessing mental health, academic performance and behavioral trends of college students using smartphones. In Proc. 2014 ACM Int. Joint Conf. on Pervasive and Ubiquitous Computing, pp. 3–14.
- 64.Liaqat D, Wu R, Gershon A, Alshaer H, Rudzicz F, de Lara E. 2018. Challenges with real-world smartwatch based audio monitoring. In Proc. 4th ACM Workshop on Wearable Systems and Applications, pp. 54–59. [DOI] [PMC free article] [PubMed]
- 65.Teague CN, Heller JA, Nevius BN, Carek AM, Mabrouk S, Garcia-Vicente F, Inan OT, Etemadi M. 2020. A wearable, multimodal sensing system to monitor knee joint health. IEEE Sensors J. 20, 10 323-10 334. ( 10.1109/JSEN.2020.2994552) [DOI] [Google Scholar]
- 66.Osmani V. 2015. Smartphones in mental health: detecting depressive and manic episodes. IEEE Pervasive Comput. 14, 10-13. ( 10.1109/MPRV.2015.54) [DOI] [Google Scholar]
- 67.Choi Y, Jeon YM, Wang L, Kim K. 2017. A biological signal-based stress monitoring framework for children using wearable devices. Sensors 17, 1936. ( 10.3390/s17091936) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Konig A, Satt A, Sorin A, Hoory R, Derreumaux A, David R, Robert PH. 2018. Use of speech analyses within a mobile application for the assessment of cognitive impairment in elderly people. Curr. Alzheimer Res. 15, 120-129. ( 10.2174/1567205014666170829111942) [DOI] [PubMed] [Google Scholar]
- 69.Hassan A, Shahin I, Alsabek MB. 2020. COVID-19 detection system using recurrent neural networks. In 2020 Int. Conf. on Communications, Computing, Cybersecurity, and Informatics (CCCI), pp. 1–5. IEEE.
- 70.Dang T, et al. 2022. Exploring longitudinal cough, breath, and voice data for COVID-19 progression prediction via sequential deep learning: model development and validation. J. Med. Internet Res. 24, e37004. ( 10.2196/37004) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Dang T, et al. 2023. Conditional neural ODE processes for individual disease progression forecasting: a case study on COVID-19. In Proc. 29th ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining, pp. 3914–3925.
- 72.Pal A, Sankarasubbu M. 2021. Pay attention to the cough: early diagnosis of COVID-19 using interpretable symptoms embeddings with cough sound signal processing. In Proc. 36th Annual ACM Symposium on Applied Computing, pp. 620–628.
- 73.Xue H, Salim FD. 2021. Exploring self-supervised representation ensembles for COVID-19 cough classification. (http://arxiv.org/abs/2105.07566).
- 74.Xu X, Nemati E, Vatanparvar K, Nathan V, Ahmed T, Rahman MM, McCaffrey D, Kuang J, Gao JA. 2021. Listen2Cough: leveraging end-to-end deep learning cough detection model to enhance lung health assessment using passively sensed audio. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 5, 1-22. ( 10.1145/3448124) [DOI] [Google Scholar]
- 75.Chamberlain D, Kodgule R, Ganelin D, Miglani V, Fletcher RR. 2016. Application of semi-supervised deep learning to lung sound analysis. In 2016 38th Annual Int. Conf. of the IEEE Engineering in Medicine and Biology Society (EMBC) pp. 804–807. IEEE. [DOI] [PubMed]
- 76.Ben-Zeev D, Scherer EA, Wang R, Xie H, Campbell AT. 2015. Next-generation psychiatric assessment: using smartphone sensors to monitor behavior and mental health. Psychiatr. Rehabil. J. 38, 218. ( 10.1037/prj0000130) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Wang R, et al. 2016. CrossCheck: toward passive sensing and detection of mental health changes in people with schizophrenia. In Proc. 2016 ACM Int. Joint Conf. on Pervasive and Ubiquitous Computing, pp. 886–897.
- 78.Jiang L, Gao B, Gu J, Chen Y, Gao Z, Ma X, Kendrick KM, Woo WL. 2018. Wearable long-term social sensing for mental wellbeing. IEEE Sens. J. 19, 8532-8542. ( 10.1109/JSEN.2018.2877427) [DOI] [Google Scholar]
- 79.Brown C, Chauhan J, Grammenos A, Han J, Hasthanasombat A, Spathis D, Xia T, Cicuta P, Mascolo C. 2020. Exploring automatic diagnosis of COVID-19 from crowdsourced respiratory sound data. (http://arxiv.org/abs/2006.05919)
- 80.Pahar M, Klopper M, Warren R, Niesler T. 2021. COVID-19 cough classification using machine learning and global smartphone recordings. Comput. Biol. Med. 135, 104572. ( 10.1016/j.compbiomed.2021.104572) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81.Alqassim S, Ganesh M, Khoja S, Zaidi M, Aloul F, Sagahyroon A. 2012. Sleep apnea monitoring using mobile phones. In 2012 IEEE 14th Int. Conf. on e-Health Networking, Applications and Services (Healthcom), pp. 443–446. IEEE.
- 82.Chen NC, Wang KC, Chu HH. 2012. Listen-to-nose: a low-cost system to record nasal symptoms in daily life. In Proc. 2012 ACM Conf. on Ubiquitous Computing, pp. 590–591.
- 83.Larson EC, Goel M, Redfield M, Boriello G, Rosenfeld M, Patel SN. 2013. Tracking lung function on any phone. In Proc. 3rd ACM Symp. on Computing for Development, pp. 1–2.
- 84.Martin A, Voix J. 2017. In-ear audio wearable: measurement of heart and breathing rates for health and safety monitoring. IEEE Trans. Biomed. Eng. 65, 1256-1263. ( 10.1109/TBME.2017.2720463) [DOI] [PubMed] [Google Scholar]
- 85.Kalantarian H, Sarrafzadeh M. 2015. Audio-based detection and evaluation of eating behavior using the smartwatch platform. Comput. Biol. Med. 65, 1-9. ( 10.1016/j.compbiomed.2015.07.013) [DOI] [PubMed] [Google Scholar]
- 86.Coppock H, Jones L, Kiskin I, Schuller B. 2021. COVID-19 detection from audio: seven grains of salt. Lancet Digit. Health 3, e537-e538. ( 10.1016/S2589-7500(21)00141-2) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 87.Wang X, Ma Y, Wang Y, Jin W, Wang X, Tang J, Jia C, Yu J. 2020. Traffic flow prediction via spatial temporal graph neural network. In Proc. Web Conf. 2020, pp. 1082–1092.
- 88.Liang Y, Ouyang K, Sun J, Wang Y, Zhang J, Zheng Y, Rosenblum D, Zimmermann R. 2021. Fine-grained urban flow prediction. In Proc. Web Conf. 2021, pp. 1833–1845.
- 89.Ben Ticha H, Absi N, Feillet D, Quilliot A. 2018. Vehicle routing problems with road-network information: state of the art. Networks 72, 393-406. ( 10.1002/net.21808) [DOI] [Google Scholar]
- 90.Ruan S, Fu X, Long C, Xiong Z, Bao J, Li R, Chen Y, Wu S, Zheng Y. 2021. Filling delivery time automatically based on couriers’ trajectories. IEEE Trans. Knowl. Data Eng. 35, 1528-1540. ( 10.1109/TKDE.2021.3100116) [DOI] [Google Scholar]
- 91.Noulas A, Scellato S, Lathia N, Mascolo C. 2012. A random walk around the city: new venue recommendation in location-based social networks. In 2012 Int. Conf. on Privacy, Security, Risk and Trust and 2012 Int. Conf. on Social Computing, pp. 144–153. IEEE.
- 92.Manzoor S, Mazhar S, Asghar A, Noor Mian A, Imran A, Crowcroft J. 2020. Leveraging mobility and content caching for proactive load balancing in heterogeneous cellular networks. Trans. Emerg. Telecommun. Technol. 31, e3739. ( 10.1002/ett.3739) [DOI] [Google Scholar]
- 93.Thorpe JR, Forchhammer BH, Maier AM. 2019. Development of a sensor-based behavioral monitoring solution to support dementia care. JMIR mHealth uHealth 7, e12013. ( 10.2196/12013) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 94.Lin Z, Lyu S, Cao H, Xu F, Wei Y, Samet H, Li Y. 2020. Healthwalks: sensing fine-grained individual health condition via mobility data. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 4, 1-26. ( 10.1145/3432229)35846237 [DOI] [Google Scholar]
- 95.Ghosh A, Rozemberczki B, Ramamoorthy S, Sarkar R. 2018. Topological signatures for fast mobility analysis. In Proc. 26th ACM SIGSPATIAL Int. Conf. on Advances in Geographic Information Systems, pp. 159–168.
- 96.Phillips JM, Pourmahmood-Aghababa H. 2020. Orientation-preserving vectorized distance between curves. (http://arxiv.org/abs/2007.15924)
- 97.Takeuchi K, Imaizumi M, Kanda S, Tabei Y, Fujii K, Yoda K, Ishihata M, Maekawa T. 2021. Fréchet kernel for trajectory data analysis. In Proc. 29th Int. Conf. on Advances in Geographic Information Systems, pp. 221–224.
- 98.Buchin K, Driemel A, Gudmundsson J, Horton M, Kostitsyna I, Löffler M, Struijs M. 2019. Approximating (k, ℓ)-center clustering for curves. In Proc. 30th Annual ACM-SIAM Symp. on Discrete Algorithms, pp. 2922–2938. Philadelphia, PA: SIAM.
- 99.Coronavirus: how they tried to curb Spanish flu pandemic in 1918. See https://www.bbc.co.uk/news/in-pictures-52564371 (accessed 21 January 2023).
- 100.Schwabe A, Persson J, Feuerriegel S. 2021. Predicting COVID-19 spread from large-scale mobility data. (http://arxiv.org/abs/2106.00356)
- 101.Chinazzi M, et al. 2020. The effect of travel restrictions on the spread of the 2019 novel coronavirus (COVID-19) outbreak. Science 368, 395-400. ( 10.1126/science.aba9757) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 102.Badr HS, Du H, Marshall M, Dong E, Squire MM, Gardner LM. 2020. Association between mobility patterns and COVID-19 transmission in the USA: a mathematical modelling study. Lancet Infect. Dis. 20, 1247-1254. ( 10.1016/S1473-3099(20)30553-3) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 103.Bonaccorsi G, et al. 2020. Economic and social consequences of human mobility restrictions under COVID-19. Proc. Natl Acad. Sci. USA 117, 15 530-15 535. ( 10.1073/pnas.2007658117) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 104.Dutta R, Gomes SN, Kalise D, Pacchiardi L. 2021. Using mobility data in the design of optimal lockdown strategies for the COVID-19 pandemic. PLoS Comput. Biol. 17, e1009236. ( 10.1371/journal.pcbi.1009236) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 105.Chang SY, et al. 2021. Supporting COVID-19 policy response with large-scale mobility-based modeling. medRxiv. ( 10.1101/2021.03.20.21254022) [DOI]
- 106.Chang S, Pierson E, Koh PW, Gerardin J, Redbird B, Grusky D, Leskovec J. 2021. Mobility network models of COVID-19 explain inequities and inform reopening. Nature 589, 82-87. ( 10.1038/s41586-020-2923-3) [DOI] [PubMed] [Google Scholar]
- 107.Wang H, Ghosh A, Ding J, Sarkar R, Gao J. 2021. Heterogeneous interventions reduce the spread of COVID-19 in simulations on real mobility data. Sci. Rep. 11, 7809. ( 10.1038/s41598-021-87034-z) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 108.Peak CM, et al. 2018. Population mobility reductions associated with travel restrictions during the Ebola epidemic in Sierra Leone: use of mobile phone data. Int. J. Epidemiol. 47, 1562-1570. ( 10.1093/ije/dyy095) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 109.Bajardi P, Poletto C, Ramasco JJ, Tizzoni M, Colizza V, Vespignani A. 2011. Human mobility networks, travel restrictions, and the global spread of 2009 H1N1 pandemic. PLoS ONE 6, e16591. ( 10.1371/journal.pone.0016591) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 110.Cebrian M. 2021. The past, present and future of digital contact tracing. Nat. Electron. 4, 2-4. ( 10.1038/s41928-020-00535-z) [DOI] [Google Scholar]
- 111.Wesolowski A, Eagle N, Tatem AJ, Smith DL, Noor AM, Snow RW, Buckee CO. 2012. Quantifying the impact of human mobility on malaria. Science 338, 267-270. ( 10.1126/science.1223467) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 112.Coughlan G, Laczó J, Hort J, Minihane AM, Hornberger M. 2018. Spatial navigation deficits—overlooked cognitive marker for preclinical Alzheimer disease? Nat. Rev. Neurol. 14, 496-506. ( 10.1038/s41582-018-0031-x) [DOI] [PubMed] [Google Scholar]
- 113.Bayat S, Mihailidis A. 2021. Outdoor life in dementia: how predictable are people with dementia in their mobility? Alzheimer’s Dement. (Amst). 13, e12187. ( 10.1002/dad2.12187) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 114.Bayat S, Babulal GM, Schindler SE, Fagan AM, Morris JC, Mihailidis A, Roe CM. 2021. GPS driving: a digital biomarker for preclinical Alzheimer disease. Alzheimer’s Res. Ther. 13, 1-9. ( 10.1186/s13195-021-00852-1) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 115.Ghosh A, Puthusseryppady V, Chan D, Mascolo C, Hornberger M. 2022. Machine learning detects altered spatial navigation features in outdoor behaviour of Alzheimer’s disease patients. Sci. Rep. 12, 3160. ( 10.1038/s41598-022-06899-w) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 116.Madan A, Cebrian M, Lazer D, Pentland A. 2010. Social sensing for epidemiological behavior change. In Proc. 12th ACM Int. Conf. on Ubiquitous Computing, pp. 291–300.
- 117.Canzian L, Musolesi M. 2015. Trajectories of depression: unobtrusive monitoring of depressive states by means of smartphone mobility traces analysis. In Proc. 2015 ACM Int. Joint Conf. on Pervasive and Ubiquitous Computing, pp. 1293–1304.
- 118.Wang Y, Zhou X, Mascolo C, Noulas A, Xie X, Liu Q. 2018. Predicting the spatio-temporal evolution of chronic diseases in population with human mobility data. In Proc. 27th Int. Joint Conf. on Artificial Intelligence, pp. 3578–3584. ( 10.24963/ijcai.2018/497) [DOI]
- 119.Zhang Y, Xu F, Li T, Kostakos V, Hui P, Li Y. 2021. Passive health monitoring using large scale mobility data. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 5, 1-23. ( 10.1145/3448078) [DOI] [Google Scholar]
- 120.Zhang Y, Xu F, Xia T, Li Y. 2021. Quantifying the causal effect of individual mobility on health status in urban space. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 5, 1-30. ( 10.1145/3494990) [DOI] [Google Scholar]
- 121.Wesolowski A, O’Meara WP, Tatem AJ, Ndege S, Eagle N, Buckee CO. 2015. Quantifying the impact of accessibility on preventive healthcare in sub-Saharan Africa using mobile phone data. Epidemiology 26, 223. ( 10.1097/EDE.0000000000000239) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 122.Huh JH, Kim TJ. 2019. A location-based mobile health care facility search system for senior citizens. J. Supercomput. 75, 1831-1848. ( 10.1007/s11227-018-2342-5) [DOI] [Google Scholar]
- 123.Burgoine T, Jones AP, Brouwer RJN, Neelon SEB. 2015. Associations between BMI and home, school and route environmental exposures estimated using GPS and GIS: do we see evidence of selective daily mobility bias in children? Int. J. Health Geogr. 14, 8. ( 10.1186/1476-072X-14-8) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 124. Indoor positioning use cases for healthcare institutions. See https://www.esri.com/arcgis-blog/products/arcgis-indoors/health/indoor-positioning-use-cases-for-healthcare/ (accessed 26 January 2023).
- 125.Hospital RTLS. See https://indoo.rs/industries/healthcare/ (accessed 26 January 2023).
- 126.Xiao F, Miao Q, Xie X, Sun L, Wang R. 2018. Indoor anti-collision alarm system based on wearable Internet of Things for smart healthcare. IEEE Commun. Mag. 56, 53-59. ( 10.1109/MCOM.2018.1700706) [DOI] [Google Scholar]
- 127.Aral S, Nicolaides C. 2017. Exercise contagion in a global social network. Nat. Commun. 8, 14753. ( 10.1038/ncomms14753) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 128.Wang Z, Derr T, Yin D, Tang J. 2017. Understanding and predicting weight loss with mobile social networking data. In Proc. 2017 ACM on Conf. on Information and Knowledge Management, pp. 1269–1278.
- 129.Serrano KJ, Coa KI, Yu M, Wolff-Hughes DL, Atienza AA. 2017. Characterizing user engagement with health app data: a data mining approach. Transl. Behav. Med. 7, 277-285. ( 10.1007/s13142-017-0508-y) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 130.McConnell MV, et al. 2017. Feasibility of obtaining measures of lifestyle from a smartphone app: the MyHeart Counts Cardiovascular Health Study. JAMA Cardiol. 2, 67-76. ( 10.1001/jamacardio.2016.4395) [DOI] [PubMed] [Google Scholar]
- 131.Mandsager K, Harb S, Cremer P, Phelan D, Nissen SE, Jaber W. 2018. Association of cardiorespiratory fitness with long-term mortality among adults undergoing exercise treadmill testing. JAMA Netw. Open 1, e183605. ( 10.1001/jamanetworkopen.2018.3605) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 132.Jones AM, Carter H. 2000. The effect of endurance training on parameters of aerobic fitness. Sports Med. 29, 373-386. ( 10.2165/00007256-200029060-00001) [DOI] [PubMed] [Google Scholar]
- 133.Savonen KP, Lakka TA, Laukkanen JA, Halonen PM, Rauramaa TH, Salonen JT, Rauramaa R. 2006. Heart rate response during exercise test and cardiovascular mortality in middle-aged men. Eur. Heart J. 27, 582-588. ( 10.1093/eurheartj/ehi708) [DOI] [PubMed] [Google Scholar]
- 134.Ellestad MH, Wan M. 1975. Predictive implications of stress testing. Follow-up of 2700 subjects after maximum treadmill stress testing. Circulation 51, 363-369. ( 10.1161/01.CIR.51.2.363) [DOI] [PubMed] [Google Scholar]
- 135.Apple. 2021 Using Apple Watch to estimate cardio fitness with VO2 max. See https://www.apple.com/healthcare/docs/site/Using_Apple_Watch_to_Estimate_Cardio_Fitness_with_VO2_max.pdf.
- 136.Spathis D, Perez-Pozuelo I, Gonzales TI, Wu Y, Brage S, Wareham N, Mascolo C. 2022. Longitudinal cardio-respiratory fitness prediction through wearables in free-living environments. npj Digit. Med. 5, 176. ( 10.1038/s41746-022-00719-1) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 137.Blalock DW, Guttag JV. 2016. Extract: strong examples from weakly-labeled sensor data. In 2016 IEEE 16th Int. Conf. Data Mining (ICDM), pp. 799–804. IEEE.
- 138.Cao ZB, Miyatake N, Higuchi M, Miyachi M, Ishikawa-Takata K, Tabata I. 2010. Predicting VO2max with an objectively measured physical activity in Japanese women. Med. Sci. Sports Exerc. 42, 179-186. ( 10.1249/MSS.0b013e3181af238d) [DOI] [PubMed] [Google Scholar]
- 139.Gonzales TI, Westgate K, Hollidge S, Lindsay T, Jeon J, Brage S. 2020. Estimating maximal oxygen consumption from heart rate response to submaximal ramped treadmill test. medRxiv. ( 10.1101/2020.02.18.20024489) [DOI]
- 140.Plasqui G, Westerterp KR. 2005. Accelerometry and heart rate as a measure of physical fitness: proof of concept. Med. Sci. Sports Exerc. 37, 872-876. ( 10.1249/01.MSS.0000161805.61893.C0) [DOI] [PubMed] [Google Scholar]
- 141.Ballinger B, et al. 2018 Deepheart: semi-supervised sequence learning for cardiovascular risk prediction. (http://arxiv.org/abs/1802.02511. )
- 142.Sarkar P, Etemad A. 2019. Self-supervised learning for ECG-based emotion recognition. (http://arxiv.org/abs/1910.07497)
- 143.Kiyasseh D, Zhu T, Clifton DA. 2020. Clocs: contrastive learning of cardiac signals. (http://arxiv.org/abs/2005.13249)
- 144.Spathis D, Perez-Pozuelo I, Brage S, Wareham NJ, Mascolo C. 2021. Self-supervised transfer learning of physiological representations from free-living wearable data. In Proc. Conf. on Health, Inference, and Learning, CHIL ’21, pp. 69–78. New York, NY: ACM.
- 145.Spathis D, Perez-Pozuelo I, Marques-Fernandez L, Mascolo C. 2022. Breaking away from labels: the promise of self-supervised machine learning in intelligent health. Patterns 3, 100410. ( 10.1016/j.patter.2021.100410) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 146.Yuan H, Chan S, Creagh AP, Tong C, Clifton DA, Doherty A. 2022. Self-supervised learning for human activity recognition using 700,000 person-days of wearable data. (http://arxiv.org/abs/2206.02909) [DOI] [PMC free article] [PubMed]
- 147.Liu X, McDuff D, Kovacs G, Galatzer-Levy I, Sunshine J, Zhan J, Poh MZ, Liao S, Di Achille P, Patel S. 2023. Large language models are few-shot health learners. (http://arxiv.org/abs/2305.15525)
- 148.Ross R, et al. 2016. Importance of assessing cardiorespiratory fitness in clinical practice: a case for fitness as a clinical vital sign: a scientific statement from the American Heart Association. Circulation 134, e653-e699. ( 10.1161/CIR.0000000000000461) [DOI] [PubMed] [Google Scholar]
- 149.Swain DP, et al. 2014. ACSM’s resource manual for guidelines for exercise testing and prescription. New York, NY: Wolters Kluwer Health/Lippincott Williams & Wilkins. [Google Scholar]
- 150.Altini M, Casale P, Penders J, Amft O. 2016. Cardiorespiratory fitness estimation in free-living using wearable sensors. Artif. Intell. Med. 68, 37-46. ( 10.1016/j.artmed.2016.02.002) [DOI] [PubMed] [Google Scholar]
- 151.Cheng H, Furnham A. 2003. Personality, self-esteem, and demographic predictions of happiness and depression. Pers. Individ. Differ. 34, 921-942. ( 10.1016/S0191-8869(02)00078-8) [DOI] [Google Scholar]
- 152.Veenhoven R. 2008. Healthy happiness: effects of happiness on physical health and the consequences for preventive health care. J. Happiness Stud. 9, 449-469. ( 10.1007/s10902-006-9042-1) [DOI] [Google Scholar]
- 153.Geukes K, Nestler S, Hutteman R, Küfner AC, Back MD. 2017. Trait personality and state variability: predicting individual differences in within-and cross-context fluctuations in affect, self-evaluations, and behavior in everyday life. J. Res. Personal. 69, 124-138. ( 10.1016/j.jrp.2016.06.003) [DOI] [Google Scholar]
- 154.Aichele S, Rabbitt P, Ghisletta P. 2016. Think fast, feel fine, live long: a 29-year study of cognition, health, and survival in middle-aged and older adults. Psychol. Sci. 27, 518-529. ( 10.1177/0956797615626906) [DOI] [PubMed] [Google Scholar]
- 155.Watson D, Clark LA, Tellegen A. 1988. Development and validation of brief measures of positive and negative affect: the PANAS scales. J. Pers. Soc. Psychol. 54, 1063. ( 10.1037/0022-3514.54.6.1063) [DOI] [PubMed] [Google Scholar]
- 156.Russell JA, Weiss A, Mendelsohn GA. 1989. Affect Grid: a single-item scale of pleasure and arousal. J. Pers. Soc. Psychol. 57, 493. ( 10.1037/0022-3514.57.3.493) [DOI] [Google Scholar]
- 157.Csikszentmihalyi M, Larson R. 1987. Validity and reliability of the experience-sampling method. J. Nerv. Ment. Dis. 175, 526-536. ( 10.1097/00005053-198709000-00004) [DOI] [PubMed] [Google Scholar]
- 158.Thieme A, Belgrave D, Doherty G. 2020. Machine learning in mental health: a systematic review of the HCI literature to support the development of effective and implementable ML systems. ACM Trans. Comput.-Hum. Interact. 27, 1-53. ( 10.1145/3398069) [DOI] [Google Scholar]
- 159.Mikelsons G, Smith M, Mehrotra A, Musolesi M. 2017. Towards deep learning models for psychological state prediction using smartphone data: challenges and opportunities. In Workshop on Machine Learning for Health (ML4H) at 31st Conf. on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017.
- 160.Taylor SA, Jaques N, Nosakhare E, Sano A, Picard R. 2017. Personalized multitask learning for predicting tomorrow’s mood, stress, and health. IEEE Trans. Affect. Comput. 11, 200-213. ( 10.1109/TAFFC.2017.2784832) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 161.Jaques N, Taylor S, Sano A, Picard R. 2017. Multimodal autoencoder: a deep learning approach to filling in missing sensor data and enabling better mood prediction. In Proc. Int. Conf. on Affective Computing and Intelligent Interaction (ACII), San Antonio, TX, USA, 23–26 October 2017.
- 162.Li B, Sano A. 2020. Extraction and interpretation of deep autoencoder-based temporal features from wearables for forecasting personalized mood, health, and stress. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 4, 1-26. ( 10.1145/3397318)35846237 [DOI] [Google Scholar]
- 163.Spathis D, Servia-Rodriguez S, Farrahi K, Mascolo C, Rentfrow J. 2019. Passive mobile sensing and psychological traits for large scale mood prediction. In Proc. 13th EAI Int. Conf. on Pervasive Computing Technologies for Healthcare, pp. 272–281.
- 164.Spathis D, Servia-Rodriguez S, Farrahi K, Mascolo C, Rentfrow J. 2019. Sequence multi-task learning to forecast mental wellbeing from sparse self-reported data. In Proc. 25th ACM SIGKDD Int. Conf. on Knowledge Discovery & Data Mining, pp. 2886–2894.
- 165.Cao B, Zheng L, Zhang C, Yu PS, Piscitello A, Zulueta J, Ajilore O, Ryan K, Leow AD. 2017. DeepMood: modeling mobile phone typing dynamics for mood detection. In Proc. 23rd ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining, pp. 747–755. New York, NY: ACM.
- 166.Sahoo PK, Thakkar HK, Lee MY. 2017. A cardiac early warning system with multi channel SCG and ECG monitoring for mobile health. Sensors 17, 711. ( 10.3390/s17040711) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 167.Wannenburg J, Malekian R, Hancke GP. 2018. Wireless capacitive-based ECG sensing for feature extraction and mobile health monitoring. IEEE Sensors J. 18, 6023-6032. ( 10.1109/JSEN.2018.2844122) [DOI] [Google Scholar]
- 168.Lopez Perales CR, Van Spall HG, Maeda S, Jimenez A, Laţcu DG, Milman A, Kirakoya-Samadoulougou F, Mamas MA, Muser D, Casado Arroyo R. 2021. Mobile health applications for the detection of atrial fibrillation: a systematic review. Europace 23, 11-28. ( 10.1093/europace/euaa139) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 169.Falter M, Budts W, Goetschalckx K, Cornelissen V, Buys R. 2019. Accuracy of Apple Watch measurements for heart rate and energy expenditure in patients with cardiovascular disease: cross-sectional study. JMIR mHealth uHealth 7, e11889. ( 10.2196/11889) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 170.Ismail S, Akram U, Siddiqi I. 2021. Heart rate tracking in photoplethysmography signals affected by motion artifacts: a review. EURASIP J. Adv. Signal Process. 2021, 1-27. ( 10.1186/s13634-020-00714-2) [DOI] [Google Scholar]
- 171.Guillodo E, et al. 2020. Clinical applications of mobile health wearable-based sleep monitoring: systematic review. JMIR mHealth uHealth 8, e10733. ( 10.2196/10733) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 172.Pandit JA, Radin JM, Quer G, Topol EJ. 2022. Smartphone apps in the COVID-19 pandemic. Nat. Biotechnol. 40, 1013-1022. ( 10.1038/s41587-022-01350-x) [DOI] [PubMed] [Google Scholar]
- 173.Toussaert S. 2021. Upping uptake of COVID contact tracing apps. Nat. Hum. Behav. 5, 183-184. ( 10.1038/s41562-021-01048-1) [DOI] [PubMed] [Google Scholar]
- 174.Salathé M. 2023. COVID-19 digital contact tracing worked—heed the lessons for future pandemics. Nature 619, 31-33. ( 10.1038/d41586-023-02130-6) [DOI] [PubMed] [Google Scholar]
- 175.Whitelaw S, Mamas MA, Topol E, Van Spall HG. 2020. Applications of digital technology in COVID-19 pandemic planning and response. Lancet Digit. Health 2, e435-e440. ( 10.1016/S2589-7500(20)30142-4) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 176.Abbas HSM, Xu X, Sun C. 2021. Role of COVIDsafe app and control measures in Australia in combating COVID-19 pandemic. Transform. Govern. People Process Policy 15, 708-719. ( 10.1108/TG-01-2021-0004) [DOI] [Google Scholar]
- 177.Lalmuanawma S, Hussain J, Chhakchhuak L. 2020. Applications of machine learning and artificial intelligence for Covid-19 (SARS-CoV-2) pandemic: a review. Chaos Solit. Fractals 139, 110059. ( 10.1016/j.chaos.2020.110059) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 178.Kucharski AJ, et al. 2020. Effectiveness of isolation, testing, contact tracing, and physical distancing on reducing transmission of SARS-CoV-2 in different settings: a mathematical modelling study. Lancet Infect. Dis. 20, 1151-1160. ( 10.1016/S1473-3099(20)30457-6) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 179.Kaptchuk G, Goldstein DG, Hargittai E, Hofman J, Redmiles EM. 2020. How good is good enough for COVID19 apps? The influence of benefits, accuracy, and privacy on willingness to adopt. (http://arxiv.org/abs/2005.04343)
- 180.Qureshi UM, Umair Z, Hancke GP. 2019. Indoor localization using wireless fidelity (WiFi) and bluetooth low energy (BLE) signals. In 2019 IEEE 28th Int. Symp. on Industrial Electronics (ISIE), pp. 2232–2237. IEEE. [DOI] [PMC free article] [PubMed]
- 181.Vuong N, Goh S, Chan S, Lau C. 2013. A mobile-health application to detect wandering patterns of elderly people in home environment. In 2013 35th Annual Int. Conf. of the IEEE Engineering in Medicine and Biology Society (EMBC), pp. 6748–6751. IEEE. [DOI] [PubMed]
- 182.Liu J, Liu H, Chen Y, Wang Y, Wang C. 2019. Wireless sensing for human activity: a survey. IEEE Commun. Surv. Tutor. 22, 1629-1645. ( 10.1109/COMST.2019.2934489) [DOI] [Google Scholar]
- 183.Yue S, Yang Y, Wang H, Rahul H, Katabi D. 2020. BodyCompass: monitoring sleep posture with wireless signals. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 4, 1-25. ( 10.1145/3397311)35846237 [DOI] [Google Scholar]
- 184.Samek W, Montavon G, Lapuschkin S, Anders CJ, Müller KR. 2021. Explaining deep neural networks and beyond: a review of methods and applications. Proc. IEEE 109, 247-278. ( 10.1109/JPROC.2021.3060483) [DOI] [Google Scholar]
- 185.Xia T, Han J, Qendro L, Dang T, Mascolo C. 2021. Uncertainty-aware COVID-19 detection from imbalanced sound data. (http://arxiv.org/abs/2104.02005)
- 186.Budd S, Robinson EC, Kainz B. 2021. A survey on active learning and human-in-the-loop deep learning for medical image analysis. Med. Image Anal. 71, 102062. ( 10.1016/j.media.2021.102062) [DOI] [PubMed] [Google Scholar]
- 187.Obermeyer Z, Emanuel EJ. 2016. Predicting the future—big data, machine learning, and clinical medicine. N. Engl. J. Med. 375, 1216. ( 10.1056/NEJMp1606181) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 188.Ramanarayanan V, Lammert AC, Rowe HP, Quatieri TF, Green JR. 2022. Speech as a biomarker: opportunities, interpretability, and challenges. Perspect. ASHA Spec. Interest Groups 7, 276-283. ( 10.1044/2021_PERSP-21-00174) [DOI] [Google Scholar]
- 189.Kathan A, et al. 2022. Personalised depression forecasting using mobile sensor data and ecological momentary assessment. Front. Digit. Health 4, 964582. ( 10.3389/fdgth.2022.964582) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 190.Armstrong J, Clifton DA. 2022. Continual learning of longitudinal health records. In 2022 IEEE-EMBS Int. Conf. on Biomedical and Health Informatics (BHI) pp. 1–6. IEEE.
- 191.Sjoding MW, Dickson RP, Iwashyna TJ, Gay SE, Valley TS. 2020. Racial bias in pulse oximetry measurement. N. Engl. J. Med. 383, 2477-2478. ( 10.1056/NEJMc2029240) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 192.Parker G, Brotchie H. 2010. Gender differences in depression. Int. Rev. Psychiatry 22, 429-436. ( 10.3109/09540261.2010.492391) [DOI] [PubMed] [Google Scholar]
- 193.Wexler J, Pushkarna M, Bolukbasi T, Wattenberg M, Viégas F, Wilson J. 2019. The what-if tool: interactive probing of machine learning models. IEEE Trans. Vis. Comput. Graph. 26, 56-65. ( 10.1109/TVCG.2019.2934619) [DOI] [PubMed] [Google Scholar]
- 194.Panigutti C, Perotti A, Panisson A, Bajardi P, Pedreschi D. 2021. FairLens: auditing black-box clinical decision support systems. Inf. Process. Manag. 58, 102657. ( 10.1016/j.ipm.2021.102657) [DOI] [Google Scholar]
- 195.Agarwal A, Dudík M, Wu ZS. 2019. Fair regression: quantitative definitions and reduction-based algorithms. In Proc. 36th Int. Conf. on Machine Learning, pp. 120–129. PMLR, vol. 97.
- 196.Yfantidou S, Constantinides M, Spathis D, Vakali A, Quercia D, Kawsar F. 2023. Beyond accuracy: a critical review of fairness in machine learning for mobile and wearable computing. (http://arxiv.org/abs/2303.15585)
- 197.Luo R, Sun L, Xia Y, Qin T, Zhang S, Poon H, Liu TY. 2022. BioGPT: generative pre-trained transformer for biomedical text generation and mining. Brief. Bioinform. 23, bbac409. ( 10.1093/bib/bbac409) [DOI] [PubMed] [Google Scholar]
- 198.Shuvo MMH, Islam SK, Cheng J, Morshed BI. 2022. Efficient acceleration of deep learning inference on resource-constrained edge devices: a review. Proc. IEEE 11, 42-91. ( 10.1109/JPROC.2022.3226481) [DOI] [Google Scholar]
- 199.Chen J, Ran X. 2019. Deep learning with edge computing: a review. Proc. IEEE 107, 1655-1674. ( 10.1109/JPROC.2019.2921977) [DOI] [Google Scholar]
- 200.Li Z, Wallace E, Shen S, Lin K, Keutzer K, Klein D, Gonzalez J. 2020. Train big, then compress: rethinking model size for efficient training and inference of transformers. In Proc. 37th Int. Conf. on Machine Learning, pp. 5958–5968. PMLR, vol. 119.
- 201.de Hond AA, et al. 2022. Guidelines and quality criteria for artificial intelligence-based prediction models in healthcare: a scoping review. npj Dig. Med. 5, 2. ( 10.1038/s41746-021-00549-7) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 202.Tangari G, Ikram M, Sentana IWB, Ijaz K, Kaafar MA, Berkovsky S. 2021. Analyzing security issues of android mobile health and medical applications. J. Am. Med. Inform. Assoc. 28, 2074-2084. ( 10.1093/jamia/ocab131) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 203.Sweeney L. 2002. k-anonymity: a model for protecting privacy. Int. J. Uncertain Fuzziness Knowl.-Based Syst. 10, 557-570. ( 10.1142/S0218488502001648) [DOI] [Google Scholar]
- 204.Liu W, Zhang Y, Yang H, Meng Q. 2023. A survey on differential privacy for medical data analysis. Ann. Data Sci. ( 10.1007/s40745-023-00475-3) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 205.Rajasekaran AS, Maria A, Rajagopal M, Lorincz J. 2022. Blockchain enabled anonymous privacy-preserving authentication scheme for internet of health things. Sensors 23, 240. ( 10.3390/s23010240) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 206.Rieke N, et al. 2020. The future of digital health with federated learning. npj Dig. Med. 3, 119. ( 10.1038/s41746-020-00323-1) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 207.Xia T, Han J, Ghosh A, Mascolo C. 2023. Cross-device federated learning for mobile health diagnostics: a first study on COVID-19 detection. In ICASSP 2023-2023 IEEE Int. Conf. on Acoustics, Speech and Signal Processing (ICASSP), pp. 1–5. IEEE.
- 208.Kotz D, Gunter CA, Kumar S, Weiner JP. 2016. Privacy and security in mobile health: a research agenda. Computer 49, 22-30. ( 10.1109/MC.2016.185) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 209.Dehling T, Gao F, Schneider S, Sunyaev A. 2015. Exploring the far side of mobile health: information security and privacy of mobile health apps on iOS and Android. JMIR mHealth uHealth 3, e3672. ( 10.2196/mhealth.3672) [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
This article has no additional data.