Skip to main content
Wiley Open Access Collection logoLink to Wiley Open Access Collection
. 2020 Nov 13;25(2):327–338. doi: 10.1002/ejp.1688

Digital manikins to self‐report pain on a smartphone: A systematic review of mobile apps

Syed Mustafa Ali 1,2,3,, Wei J Lau 4, John McBeth 1,2,3, William G Dixon 1,2,3, Sabine N van der Veer 1,2,3,4
PMCID: PMC7839759  PMID: 33113241

Abstract

Background

Chronic pain is the leading cause of disability. Improving our understanding of pain occurrence and treatment effectiveness requires robust methods to measure pain at scale. Smartphone‐based pain manikins are human‐shaped figures to self‐report location‐specific aspects of pain on people's personal mobile devices.

Methods

We searched the main app stores to explore the current state of smartphone‐based pain manikins and to formulate recommendations to guide their development in the future.

Results

The search yielded 3,938 apps. Twenty‐eight incorporated a pain manikin and were included in the analysis. For all apps, it was unclear whether they had been tested and had end‐user involvement in the development. Pain intensity and quality could be recorded in 28 and 13 apps, respectively, but this was location specific in only 11 and 4. Most manikins had two or more views (n = 21) and enabled users to shade or select body areas to record pain location (n = 17). Seven apps allowed personalising the manikin appearance. Twelve apps calculated at least one metric to summarise manikin reports quantitatively. Twenty‐two apps had an archive of historical manikin reports; only eight offered feedback summarising manikin reports over time.

Conclusions

Several publically available apps incorporated a manikin for pain reporting, but only few enabled recording of location‐specific pain aspects, calculating manikin‐derived quantitative scores, or generating summary feedback. For smartphone‐based manikins to become adopted more widely, future developments should harness manikins’ digital nature and include robust validation studies. Involving end users in the development may increase manikins’ acceptability as a tool to self‐report pain.

Significance

This review identified and characterised 28 smartphone apps that included a pain manikin (i.e. pain drawings) as a novel approach to measure pain in large populations. Only few enabled recording of location‐specific pain aspects, calculating quantitative scores based on manikin reports, or generating manikin feedback. For smartphone‐based manikins to become adopted more widely, future studies should harness the digital nature of manikins, and establish the measurement properties of manikins. Furthermore, we believe that involving end users in the development process will increase acceptability of manikins as a tool for self‐reporting pain.

1. INTRODUCTION

Chronic pain is a common cause of disability worldwide (Vos et al., 2012) with significant economic costs for society (Breivik et al., 2013). For example, in the United Kingdom, an estimated 28 million adults live with chronic pain (Fayaz et al., 2016), many of them with pain in multiple sites (Carnes et al., 2007). In addition to its high prevalence, chronic pain also has a substantial impact on the physical well‐being and mental health of those suffering from it (Breivik et al., 2006; Hadi et al., 2019; Tsang et al., 2008), which in turn negatively affects other outcomes (Lacey et al., 2015; Smith et al., 2018).

The clinical and self‐management of chronic pain is suboptimal (Mills et al., 2016), and there is a recognised need for epidemiological studies in order to improve this (Dorner, 2018). However, the complex and fluctuating nature of pain (Diatchenko et al., 2013); the fact that people from different backgrounds may express and report their pain differently (Campbell & Edwards, 2012; Krupić et al., 2019); and a lack of high‐quality population studies (Fayaz et al., 2016) hampers our understanding of the pain occurrence and progression and pain mechanisms, as well as the development of effective therapeutical interventions.

Addressing these challenges requires reliable and valid measurement methods (Erdek & Pronovost, 2004; Fayaz et al., 2016) that can be used longitudinally. Currently, there are a range of instruments available to measure pain, ranging from single‐item measures of pain location or overall pain intensity (e.g. Visual Analogue Scale) to multi‐item questionnaires covering several pain domains (e.g. the McGill Pain Questionnaire) (McCormick & Frampton, 2019; Pogatzki‐Zahn et al., 2019). However, whereas single items tend to oversimplify the pain experience (Younger et al., 2009), more complex instruments often require high levels of literacy and engagement (Wood, 2004).

Pain manikins, also known as pain drawings or pain diagrams, are human‐shaped figures where people can shade areas to self‐report the location of their pain more accurately (Jang et al., 2014). Owing to their reliability (Southerst et al., 2013), pain manikins have been used widely to assess pain and support better chronic pain management (Baeyer et al., 2011; Cruder et al., 2018; Grunnesjö et al., 2006). A recent systematic literature review by Shaballout and colleagues provided a comprehensive overview of methodological milestones in the development of pain manikins to date. They showed that paper‐based manikins first appeared in 1949, with a version for personal and tablet computers following 40 years later (Shaballout et al., 2019).

Due to their high uptake and integration into people's daily life, smartphones are an increasingly common technology for collecting self‐reported health data for research (Hicks et al., 2019), self‐management (Najm et al., 2019) and clinical care (Rowland et al., 2020). Compared to their currently available pc and tablet‐based counterparts, smartphone‐based pain manikins require a novel approach to allow for smaller and non‐standardised screen sizes. A smartphone‐based pain manikin opens up the possibility of collecting manikin reports at scale and pace via personal mobile devices. The 2019 app review by Zhao et al. reported that 11 of the 36 included pain management apps incorporated some kind of body map or manikin (Zhao et al., 2019). Yet, Shaballout et al. did not identify the use of smartphones for pain manikin reports as a methodological milestone. Furthermore, Zhao's review did not provide much detail on the features of the body map functionality, leaving it largely unknown to what extent smartphone‐based manikins are available for measuring pain.

This review, therefore, aimed to gain insight into the current state of play for smartphone‐based pain manikins and formulate recommendations to guide future developments. Specific objectives were to (a) Identify and characterise apps that include a pain manikin; and (b) Within these apps, describe the features of the manikin functionality. We addressed these objectives through a systematic review of app stores. We expect our review to contribute to the development and adoption of manikins for improved pain measurement.

2. METHODS

We used the Preferred Reporting Items for Systematic Reviews and Meta‐Analyses (PRISMA) guidelines (Moher et al., 2009) to report the results of our app review process, where relevant.

2.1. Search strategy

We searched Apple's App store (UK) and the Google Play store in April 2019 using a two‐step search strategy. First, we performed four separate searches using ‘pain’, ‘pain tracker’, ‘pain diary’ and ‘pain diagram’ as search terms. From these initial searches, we identified painful conditions for which apps were frequently available, such as rheumatoid arthritis, migraine and headache. In the second step, we then used these painful conditions as search terms in combination with the terms from the first search.

To ensure the comprehensiveness of our results, we complemented our search strategy by hand‐searching apps included in reviews published in the last 5 years that focused on pain management (Devan et al., 2019; Lalloo et al., 2017; Zhao et al., 2019).

2.2. Inclusion criteria

Apps were eligible for inclusion if they:

  • Included a manikin that allowed people to self‐report their pain. For the purpose of this review, we defined a manikin as “a line drawn to outline the human body and to allow reporting of pain in any particular part or location of the body” (Schierhout & Myers, 1996), while also including three‐dimensional body shapes. We excluded non‐interactive manikins, i.e. manikins that did not allow users to manipulate or draw on the body shape (e.g. manikins included as an illustration for educational purposes);

  • Had people affected by pain as their target users; apps aimed primarily at healthcare professionals were excluded;

  • Had a title and description in English;

  • Were available in the United Kingdom;

  • Published at any date.

2.3. App selection process

After de‐duplicating the results from the searches, two reviewers (SMA, WJL) independently screened all apps for relevance using the criteria as stated above, based on the title and description as provided in the app store. All apps (both free and paid) deemed relevant were downloaded for full review to confirm eligibility; for apps that were excluded at this stage we recorded the reason why. Disagreements about inclusion between reviewers were resolved through discussion, involving a third reviewer (SNVDV) where needed.

2.4. Data extraction and synthesis

Each included app was downloaded on a Samsung Galaxy S6 (Android Version 7.0) or iPhone X (IOS 11.2.1) for duplicate and independent data extraction by two reviewers (SMA, WJL). We developed a data extraction form in Microsoft Excel, informed by the published manikin literature (Barbero et al., 2015; Bryner, 1994; Cruder et al., 2018; Hüllemann et al., 2017; Jamison et al., 2004; Leoni et al., 2017; Neubert et al., 2018; Shaballout et al., 2019; Southerst et al., 2013; Stinson et al., 2014; Wenngren & Stålnacke, 2009; Zhao et al., 2019), the Mobile Application Rating Scale (MARS) (Stoyanov et al., 2015) and work from the Initiative on Methods, Measurement, and Pain Assessment in Clinical Trials (IMMPACT) (Dworkin et al., 2005). The research team piloted the data extraction form for completeness and clarity. The final version included the following categories of items (see Tables S1 and S2 in Supplement for the complete list of items and their descriptions):

  • General app characteristics, including: app name; developer; geographical region; availability (App store; Google Play; both); affiliation (i.e. professional or professional organisations/associations involved in app development yes; no); if end users had been involved in the development process (yes; no/unknown); purchase costs (free; in‐app purchases; costs at download); and log‐in requirements (yes; no);

  • App quality, using MARS (Stoyanov et al., 2015). MARS consists of 19 items that are rated on a 5‐point scale (1 = Inadequate, 2 = Poor, 3 = Acceptable, 4 = Good and 5 = Excellent). Items are grouped into four dimensions: engagement (e.g. interactivity – does it allow user input, provide feedback and contain prompts?); functionality (e.g. ease of use – how easy is it to learn how to use the app?); aesthetics (e.g. visual appeal – how good does the app look?) and information (e.g. credibility – does the app come from a legitimate source?). In line with previous studies (Devan et al., 2019; Salazar et al., 2018; Stoyanov et al., 2015), two reviewers independently rated for each app all MARS items and calculated mean scores per dimension and the mean overall quality score across dimensions. Discrepancies in dimension and overall scores between reviewers were resolved by taking the mean. To further ensure a shared understanding of MARS items between reviewers, we provided additional information to illustrate how to apply each item to the pain manikin apps in our review. For the MARS item “Has the app been trialled and tested,” we used the app name to search Google Scholar for empirical studies evaluating the usability, effectiveness or other aspects of the app.

  • Recording of pain intensity and quality; this category referred whether users could record pain intensity and pain quality in the app, and to what extent it was possible to enter different values for different locations. Pain intensity and pain quality are recommended as essential measures within the core outcome domain of ‘pain’, which is one of the six domain to be considered when designing chronic pain clinical trials, as recommended by the Initiative on Methods, Measurements, and Pain Assessment in Clinical Trials (IMMPACT) (Dworkin et al., 2005). IMMPACT describes pain intensity as the overall magnitude of the pain, while referring to pain quality as capturing its sensory and affective characteristics. We considered other pain aspects recommended by IMMPACT (e.g. rescue treatment) out of scope for this review because they were less likely to be location specific.

  • Manikin features, including items related to: layout (i.e. the look of the manikin; e.g. level of detail of the body figure); interaction (i.e. the feel of the manikin; e.g. method for recording pain location in the manikin) and any other input directly related to manikin (e.g. free‐text field qualitatively describing the manikin report) (see Table S1 for additional detail on individual items)

  • Summaries and feedback of manikin reports; this referred to if and how manikin reports were summarised (i.e. manikin metrics, e.g. number of painful areas, mean pain intensity score across locations); how manikin reports and summaries were presented back to users; and whether users could share reports and summaries with others.

Reviewers met regularly throughout the data extraction process to develop a common understanding of items and solve any discrepancies, involving a third reviewer (SNVDV) if needed. We thematically analysed and synthesised the extracted data in line with our study objectives, using descriptive summary statistics where relevant.

3. RESULTS

Figure 1 shows how our search yielded 3,938 apps, of which we deemed 31 eligible for inclusion in our review. However, since three apps became unavailable during the review process, we finally included 28 apps for data extraction and synthesis.

Figure 1.

Figure 1

flow diagram of the app selection process

Table 1 displays the general characteristics of included apps. More than half of them (n = 15) were developed in North America, and 12 were available in the Apple app store either exclusively or together with Google App store. The majority of apps were available for download at no cost (n = 24), with four offering additional functionality after in‐app purchases. Of the four requiring payment to download, costs ranged from US$ 0.99 to 6.99. Nine apps mentioned the input from healthcare professionals or professional associations (e.g. the International Children's Palliative Care Network, Society of Critical Care Medicine) in the app description. However, none of the app descriptions mentioned the involvement of end users in the process of app development, leaving this aspect unknown.

TABLE 1.

General characteristics and quality assessment scores of included apps (n = 28) (values are numbers (%) unless indicated otherwise)

Characteristic Number (%)
General characteristics
Geographical region
North America 15 (54)
Europe 8 (29)
Asia 2 (7)
Africa 2 (7)
Australia 1 (3)
Availability
Apple App store only 12 (43)
Google Play only 6 (21)
Both 10 (36)
Affiliation a
Yes 9 (32)
No 19 (68)
Purchase costs
Free to download 20 (72)
Free to download with in‐app purchases 4 (14)
Costs to download 4 (14)
Login requirements
Yes 13 (46)
No 15 (54)
App quality (MARS scores) b Mean (SD)
Overall quality score 3.3 (0.8)
Engagement dimension score 3.0 (0.9)
Functionality dimension score 3.8 (0.8)
Aesthetic dimension score 3.3 (0.9)
Information quality dimension score 2.9 (0.8)

Abbreviations: MARS: Mobile App Rating Scale; SD: Standard Deviation.

a

Refers to mention of involvement of individual health professionals and/or professional institution/association in the app description

b

Scores can range from 1 (inadequate) to 5 (excellent).

3.1. App quality

Table 1 shows that, on a scale of 1 to 5 (with higher scores indicating better quality), the mean overall quality score across apps was 3.3 (SD = 0.8); the best‐ and worst‐rated app scoring 4.5 and 2.1 respectively. Mean scores for the engagement, functionality, aesthetics and information dimensions were 3.0 (standard deviation [SD], 0.9), 3.8 (SD, 0.8), 3.3 (SD, 0.9) and 2.9 (SD, 0.8), respectively (see supplementary Table S3 for app‐specific scores). Most apps (n = 21) scored highest on the functionality dimension and lowest on the information dimension (n = 18). Based on the app description and our Google Scholar search, none of the assessed apps appeared to have been trialled or tested (e.g. for usability, reliability, effectiveness).

3.2. Recording of pain intensity and pain quality

Table 2 shows that pain intensity could be recorded in all 28 apps, but only in 11 could users enter different intensity scores for different locations. Thirteen apps enabled recording of pain quality, with four giving the option of location‐specific values. Examples of available pain quality descriptors included ‘stabbing’, ‘burning’, ‘throbbing’ and ‘shooting’. The four apps with location‐specific pain quality also enabled recording of location‐specific intensity.

TABLE 2.

Overview of which apps enabled self‐reporting of pain intensity and pain quality a , and which of those allowed recording of location‐specific pain intensity and pain quality

App name ↓ Pain intensity Pain quality

Self‐report enabled

Location specific

Self‐report enabled

Location specific
Chronic Pain tracker Yes Yes Yes
Feel My Pain Yes Yes Yes Yes
GeoPain @Home Yes Yes
iMigraine – Migraine Tracker Yes
Joint Pain Tracker Yes
Mend+ Yes Yes
Migraine Buddy – The Migraine and Headache Tracker Yes
Migraine Headache Diary Head App Yes Yes
Migraine Monitor Yes
My Pain Diary: GOLD EDITION Yes Yes
myVectra Yes Yes
Ouchie: a pain management companion Yes Yes
Pain Assessment Tool for Children Yes Yes Yes
Pain Cal Yes Yes
Pain Companion Yes Yes
Pain Diary (Privacy Friendly) Yes Yes
Pain Diary & Forum CatchMyPain Yes Yes Yes
Pain & Opioid Safety Yes Yes Yes
Pain Scored Yes Yes
Pain Tracker & Diary Yes Yes Yes Yes
Pain Tracker HD Yes
Pain Tracker Yes
Patient Communicator Yes Yes Yes Yes
RA Monitor Yes
RAPA – RA Patient Application Yes
RheumBuddy Yes
Simple Pain Scale Yes
The O Lab Pain Tracking Yes Yes
a

IMMPACT (Dworkin et al., 2005) describes pain intensity as the overall magnitude of the pain, and pain quality as the different sensory and affective characteristics of pain.

3.3. Manikin features

Table 3 shows that the majority of manikins: had more than one view (n = 21; e.g. front and back); used labels to support orientation (n = 15; mostly ‘left’ and ‘right’) and represented the body figure two dimensionally (n = 19) with at least moderately detailed body anchors (n = 16). Figure 2 displays six examples of included pain manikins to illustrate the broad range of manikin features and how we assessed them.

TABLE 3.

Summary of manikin features

Manikin features Number (%)
Number of views
One 7 (25)
Two 15 (54)
More than two 6 (21)
Labels for orientation
Yes 15 (54)
No 13 (46)
Number of dimensions
Two 19 (68)
Three 9 (32)
Level of detail of body anchors a
Low 11 (39)
Moderate 9 (32)
High 8 (29)
Zoom option
Yes 13 (46)
No 15 (54)
Manikin personalisation options
Yes 7 (25)
No 21 (75)
Free‐text field directly related to manikin report
Yes 5 (18)
No 23 (82)
Method to record pain location
Shading any area by drawing on manikin b 8 (28)
Shading pre‐defined areas by tapping on manikin c 9 (32)
Marking area by tapping on manikin d 10 (36)
Selecting locations from pre‐defined list/labels e 1 (4)
Option to undo last drawing action
Yes 23 (82)
No 5 (18)
a

For example, anchors indicating clavicle, abdominal muscles, scapulae, spine, hips, thighs, knees, ankle and groin.

b

See panels a, c, d in Figure 2

c

See panel b in Figure 2

d

Refers to apps where users could tap on the body figure to indicate location, after which a mark appeared (e.g. cross, dot or label). See panel e in Figure 2

e

See panel f in Figure 2

Figure 2.

Figure 2

Examples of manikins in included apps to illustrate the variety of manikin features and how we assessed them. GeoPain @Home (a): fully rotatable 3‐D manikin with highly detailed body anchors; zoom and undo options; user can add free‐text description to manikin report. Pain Diary (b): 2‐D manikin with two views, front‐back labels and no body anchors; users can tap on manikin to select segments with only one level of pain intensity. Pain Diary & Forum CatchMyPain (c): 3‐D manikin, offering four views, with high level of body anchors; users can shade areas by drawing to record location‐specific pain intensity. Pain Tracker & Diary (d): two‐dimensional manikin, with gender personalisation and left/right labels; moderate level of body anchors; users can shade areas by drawing on manikin with location‐specific pain intensity and quality (shown as layers); zoom and undo options; users can add free‐text description to manikin reports. Patient Communicator (e): 3‐D manikin, offering two views, with high level of anchor detail and gender personalisation. Users can tap on manikin to mark pain location with a circle and can assign location‐specific pain intensity score and can chose pain quality from a pre‐defined list; a number represents location‐specific intensity; undo option. Pain Tracker (f): 3‐D manikin with 1 view and a high level of anchor detail; users can select a location from a drop‐down list with pre‐defined locations or tapping the label for that location on the manikin; no recording of location‐specific pain aspects

In 13 apps, users could zoom in on specific areas of the manikin. The GeoPain @Home app (see Figure 2, panel a) had a fully rotatable manikin, offering users an infinite number of views. Some apps (n = 7) allowed personalisation of the body diagram. For example, in Pain Assessment Tool for Children app, users could select the gender, age and skin colour of the manikin. Moreover, in five apps, users could add free‐text notes directly related to their manikin report.

The most common method that apps offered for indicating pain location was by tapping on predefined segments on the manikin in order to shade them (n = 13; e.g. Figure 2, panel b). In more than one quarter of the apps (n = 8), users could shade any painful area by drawing directly on the manikin (e.g. Figure 2, panel a, c, d), while the remaining apps used simpler methods, including marking a location with a cross, dot or other label by tapping on manikin (n = 5; e.g. Figure 2, panel e), and selecting pre‐defined labels on the manikin or from a list (n = 2; e.g. Figure 2, panel f).

3.4. Summaries and feedback of manikin reports

Table 4 shows that a total of 12 apps calculated 9 different manikin metrics to summarise manikin reports. The most common metrics were related to location (n = 6), such as the frequency with which a body manikin area was affected in a specified time period. Five apps calculated metrics related to pain extent. Of those, only three calculated the % of body area affected, despite eight apps having the theoretical possibility to provide this metric (i.e. because users could shade any painful area directly on the manikin). Of the 14 apps that enabled users to record their location‐specific pain intensity, four provided mean intensity as a metric. Only two apps combined multiple pain aspects into a single, composite metric: GeoPain @Home and Feel my Pain calculated pain scores by combining pain extent and intensity, but without specifying the score's underlying formula.

TABLE 4.

Overview quantitative manikin metrics that apps provided to summarise manikin reports; in total, 12 apps provided at least one manikin metric

Manikin metric Metric definition Apps that calculated the metric
Pain extent (n = 5)
% body manikin area affected % of the total manikin that had been shaded with any intensity GeoPain @ Home a ; Pain Tracker & Diary; Pain Diary & Forum CatchMyPain
Number of affected body manikin areas Total number of pre‐defined body manikin segments or joints that had been selected with any intensity Pain Scored; Pain Diary (Privacy Friendly)
Pain location (n = 6)
Body manikin area affected in a specified time period Whether a pre‐defined body manikin segment or joint had been selected with any intensity yes/no in a time period myVectra; RAPA – RA Patient Application
Frequency with which a body manikin area was affected in a specified time period Number of times a pre‐defined body manikin segment or joint had been selected with any intensity in a time period Pain Scored d ; Migraine Headache Diary HeadApp; RheumaBuddy;
Number of symptoms per affected body manikin area Number of symptoms that users had selected for a pre‐defined body manikin segment RA Monitor;
Pain intensity (n = 4)
Mean intensity across the body manikin Weighted mean pain intensity across all shaded areas on the manikin b GeoPain @ Home a ; Pain Tracker & Diary;
Mean intensity per affected body manikin area Mean pain intensity for a specific body manikin area or pre‐defined segment

GeoPain @ Home a , c ; Pain Diary & Forum CatchMyPain; The O Lab Pain Tracking

Composite metrics (n = 2)
P.A.I.N.S. score Combines mean intensity and % body area affected, but the underlying formula is not specified GeoPain @Home a
Pain score e Not provided Feel my Pain
a

GeoPain @Home provided all metrics at the level of the full body manikin, as well as for 14 pre‐defined body parts (e.g. mouth, right hand, pelvis, etc.)

b

We assumed that apps calculated a weighted mean intensity, but only Pain Tracker & Diary explicitly stated this.

c

Displayed as colours in a summary manikin; colours referred to the original pain intensity scale of 0–10

d

Displayed as colours in a summary manikin; colours referred to three categories (low (<−1σ), medium (σ) and high (>1σ)). No explanation of what σ means in this context.

e

Presented as ‘'Your score x out of 10', with higher values reflecting a better state

The majority of apps (n = 22) had an archive function in the form of a list, calendar or graph where users could access historical, individual entries. For the majority of these (n = 15), entry‐level feedback included an image of the manikin report. In four apps, users could easily browse through the archive of images (e.g. using a slider) for a ‘pain time‐lapse’ to see how their pain had changed over a certain period. Only some apps (n = 8) provided feedback in the form of a summary manikin that combined pain entries across manikin reports: for example, by shading all body segments that had been selected at least once over a specified time period (see Figure 3, panel a), or by creating a ‘heat map’ to feed back the mean location‐specific intensity (see Figure 3, panel c). Of the 12 apps that calculated at least one manikin metric, five provided users with a graph of that metric over time (see Figure 3, panel b). If fed back numerically, metric values were presented at the level of one manikin report (n = 6), across reports in a specified time period (n = 8), or both (n = 1). Geopain @Home also presented a change in metric value between the first and last manikin report in the selected time period.

Figure 3.

Figure 3

Examples of different types of manikin feedback. Pain Scored (a): body segments are shaded based on the selection of a body segment at least once over a specified time period. Migraine Headache Diary Head app (b): single metric over time, presented as a graph. Geo Pain (c): mean location‐specific intensity, presented as a heat map

Eleven apps enabled users to share their manikin report with others via email, e.g. by creating a PDF, or a URL to access the report directly. Among these 11 apps, Ouchie also allowed users to share their reports with an in‐app community to facilitate discussion and peer support.

4. DISCUSSION

This review included 28 smartphone apps that were publically available in app stores in the United Kingdom and that incorporated a manikin for pain reporting. The mean overall quality score of apps was 3.3 of 5. For all of them, it was unclear if they had been trialled or tested, and whether they had involved end users in the development. All apps enabled users to record pain intensity, but this was location specific in less than half of them. The majority of manikins had at least two or more views presented two dimensionally and with at least moderately detailed body anchors, where users could select pre‐defined body segments to indicate the location of their pain. One in four apps offered the option of personalising the manikin appearance. Almost one in two apps calculated at least one metric to quantitatively summarise manikin reports, with location being most common. Although almost all apps had an archive function where users could access historical manikin reports, only few offered feedback that summarised manikin reports over time.

4.1. Relation to other studies

We identified 31 apps that incorporated a pain manikin, of which we included 28 in our review. This number is similar to or higher than other recent reviews of pain management apps, which identified 36 (of which 11 included a manikin) (Zhao et al., 2019), 19 (Devan et al., 2019) and 10 (Lalloo et al., 2017). All expressed concerns about ad hoc (Zhao et al., 2019) or no engagement (Devan et al., 2019; Lalloo et al., 2017) of end users in the process of app development; our findings seem to support these concerns.

Only a few apps in our review offered personalisation of manikin features. This finding aligns with that of Devan et al, who reported that none of the self‐management apps included culturally tailored aspects (Devan et al., 2019). A meta‐ethnographic review exploring patients’ perceptions, beliefs and experience of mHealth apps highlighted the lack of personalisation as one of the major concerns (Vo et al., 2019).

We found that 11 of 28 apps enabled users to report location‐specific pain aspects in the manikin, which is a similar proportion identified by Zhao et al. (2019), who reported this for 4 of 11 apps included in their review.

Our results also showed that none of the apps, or the manikins within them, were trialled or tested. This hampers their use in clinical or research settings, which is different for paper‐based manikins, which have been evaluated more extensively (Chatterton et al., 2013; Southerst, et al., 2013). Zhao et al. and Devan et al. similarly concluded that the pain management apps in their review were not suitable for clinical usage, mainly due lack of: patient and clinician involvement in the development process; support for patients to communicate with clinicians and evaluations of effectiveness (Devan et al., 2019; Zhao et al., 2019).

4.2. Implications for future development of smartphone‐based pain manikins

4.2.1. Developers should harness the digital nature of manikins

For smartphone‐based manikins to be adopted more widely, they need to offer clear advantages compared to their well‐established, paper‐based counterparts (O’Donnell & Curley, 1985; Öhlund et al., 1996). Their digital nature has several potential benefits, including: easy, longitudinal recording of location and location‐specific aspects of pain; more realistic and dynamic features (e.g. a three‐dimensional manikin that users can view from any angle); quick and reliable processing of manikin reports into quantitative pain scores (i.e. manikin metrics), established (e.g. % body area affected) as well as novel ones (e.g. composite metric combining pain extent and intensity); and instant and customisable feedback of manikin reports over time. However, our review shows that these benefits have not yet fully materialised, with most manikins offering relatively basic functionality. This suggests there is ample opportunity for developers to better harness the digital nature of future smartphone‐based manikins, and for researchers to investigate the effect of these enhancements on users’ engagement with manikins over time.

4.3. Researchers should establish the measurement properties of manikins

Confirming the robustness of smartphone‐based manikins as a pain measurement instrument is another prerequisite for these tools to be adopted more widely. In keeping with previous reviews (De La Vega & Miró, 2014; Haskins et al., 2017), our review suggests that there is a gap between apps that are publically available and those that have been investigated in research studies. There have been studies on the reliability of pain manikins developed for larger mobile devices (e.g. tablets) (Barbero et al., 2015; Leoni et al., 2017; Neubert et al., 2018), but those findings do not necessarily translate to manikins developed for smaller, non‐standardised smartphone screens. Future research should, therefore, focus on assessing the measurement properties (e.g. reliability, validity, responsiveness) of smartphone‐based pain manikins in line with international guidance (Mokkink et al., 2010), in order for them to become an accepted standard for measuring location‐specific pain aspects in trials and epidemiological studies. This will also contribute to embedding these manikins into interventions aimed at improving clinical and self‐management of painful conditions.

4.4. Developers should consider involving representative end user groups in the development process

Involving end users will ensure that apps incorporating a manikin can reach their intended objectives (McCurdie et al., 2012). Furthermore, we believe that end user involvement will increase the chance of manikin reports being an adequate reflection of people's pain, thereby increasing their acceptability as a self‐reporting tool. Pain prevalence is highest among black and Asian minority ethnicities (Macfarlane et al., 2015), who may experience or report pain differently compared to Caucasians (Mills et al., 2019). The same has been suggested for other groups, such as older people, or those suffering from multisite pain (Mills et al., 2019). Therefore, involving representative end user groups may ultimately contribute to better content and cross‐cultural validity of manikins as measurement instruments (Mokkink et al., 2010), and enhance the usability and usefulness of manikins for a broad range of people living with a painful condition. The fact that only few of the apps allowed cultural or other personalisation of the manikin appearance may indicate that there is room for improvement in this area.

4.5. Limitations

One limitation of our review is that it was designed to identify manikins that were incorporated within apps that were publically available in UK app stores. This implies that our search strategy did not allow us to identify smartphone‐based manikins offered in app stores outside the United Kingdom, or by healthcare providers or researchers via other routes. However, the literature review by Shaballout et al. (Shaballout, et al., 2019) did not identify them as a methodological milestone, suggesting that peer‐reviewed reports on smartphone‐based manikins are scarce at most. This means there is still a gap in knowledge on how screen size and other device characteristics affect user engagement in and measurement properties of smartphone‐based manikins.

Another limitation is that for our data on general app characteristics and parts of the app quality appraisal, we relied on information provided in the app stores, but without contacting the app developers to confirm and complement it. This would have allowed us to better interpret missing information on, for example, end user involvement (i.e. distinguish ‘not reported’ from ‘not undertaken’) and whether apps had been trialled.

5. CONCLUSION

This systematic review of app stores identified a substantial number of publically available apps that incorporated a manikin for pain reporting. Although this number is encouraging, only few apps offered functionality to: record location‐specific pain aspects; calculate quantitative scores based on manikin reports or generate engaging manikin summary feedback. This limits their added value compared to their paper‐based counterparts. Future developments of smartphone‐based manikins should better harness the manikins’ digital nature and include robust studies to establish their measurement properties. We also believe that involving end users in the development process will increase acceptability of manikins as a tool for self‐reporting pain. Ultimately, this will lead to smartphone‐based manikins becoming an accepted standard for measuring location‐specific pain aspects for research, self‐management and clinical care.

CONFLICT OF INTEREST

None of the authors has any conflict of interest.

DISCLOSURE

This work, either in part or full, is not presented previously.

AUTHOR CONTRIBUTIONS

SNVDV, WGD and JM conceived the idea of this app review. SMA, WJL and SNVDV extracted and synthesised the data. All authors were involved in interpreting the data. SMA drafted the manuscript. All authors revised it critically for important intellectual content and approved the final version of manuscript for submission. SMA and SNVDV take responsibility for the integrity of the work as a whole.

Supporting information

Table S1‐S3

ACKNOWLEDGEMENT

None of the authors have conflict of interest to declare.

Ali SM, Lau WJ, McBeth J, Dixon WG, van der Veer SN. Digital manikins to self‐report pain on a smartphone: A systematic review of mobile apps. Eur J Pain. 2021;25:327–338. 10.1002/ejp.1688

Funding information

The study was supported by the Centre for Epidemiology Versus Arthritis (grant number 21755).

REFERENCES

  1. Barbero, M. , Moresi, F. , Leoni, D. , Gatti, R. , Egloff, M. , & Falla, D. (2015). Test‐retest reliability of pain extent and pain location using a novel method for pain drawing analysis. European Journal of Pain (United Kingdom), 19(8), 1129–1138. 10.1002/ejp.636 [DOI] [PubMed] [Google Scholar]
  2. Breivik, H. , Collett, B. , Ventafridda, V. , Cohen, R. , & Gallacher, D. (2006). Survey of chronic pain in Europe: Prevalence, impact on daily life, and treatment. European Journal of Pain, 10(4), 287 10.1016/j.ejpain.2005.06.009 [DOI] [PubMed] [Google Scholar]
  3. Breivik, H. , Eisenberg, E. , & O’Brien, T. (2013). The individual and societal burden of chronic pain in Europe: The case for strategic prioritisation and action to improve knowledge and availability of appropriate care. BMC Public Health, 13(1), 1229 10.1186/1471-2458-13-1229 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Bryner, P. (1994). Extent measurement in localised low‐back pain: A comparison of four methods. Pain, 59(2), 281–285. 10.1016/0304-3959(94)90081-7 [DOI] [PubMed] [Google Scholar]
  5. Campbell, C. M. , & Edwards, R. R. (2012). Ethnic differences in pain and pain management. Pain Management, 2(3), 219–230. 10.2217/pmt.12.7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Carnes, D. , Parsons, S. , Ashby, D. , Breen, A. , Foster, N. E. , Pincus, T. , Vogel, S. , & Underwood, M. (2007). Chronic musculoskeletal pain rarely presents in a single body site: Results from a UK population study. Rheumatology, 46(7), 1168–1170. 10.1093/rheumatology/kem118 [DOI] [PubMed] [Google Scholar]
  7. Chatterton, B. D. , Muller, S. , Thomas, M. J. , Menz, H. B. , Rome, K. , & Roddy, E. (2013). Inter and intra‐rater repeatability of the scoring of foot pain drawings. Journal of Foot and Ankle Research, 6(1), 44 10.1186/1757-1146-6-44 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Cruder, C. , Falla, D. , Mangili, F. , Azzimonti, L. , Araújo, L. S. , Williamon, A. , & Barbero, M. (2018). Profiling the location and extent of musicians’ pain using digital pain drawings. Pain Practice, 18(1), 53–66. 10.1111/papr.12581 [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. De La Vega, R. , & Miró, J. (2014). mHealth: A strategic field without a solid scientific soul. A systematic review of pain‐related apps. PLoS One, 9(7), 10.1371/journal.pone.0101312 [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Devan, H. , Farmery, D. , Peebles, L. , & Grainger, R. (2019). Evaluation of self‐management support functions in apps for people with persistent pain: Systematic review. JMIR MHealth and UHealth, 7(2), e13080 10.2196/13080 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Diatchenko, L. , Fillingim, R. B. , Smith, S. B. , & Maixner, W. (2013). The phenotypic and genetic signatures of common musculoskeletal pain conditions. Nature Reviews Rheumatology, 9(6), 340–350. 10.1038/nrrheum.2013.43 [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Dorner, T. E. (2018). Pain and chronic pain epidemiology. Wiener Klinische Wochenschrift, 130(1–2), 1–3. 10.1007/s00508-017-1301-0 [DOI] [PubMed] [Google Scholar]
  13. Dworkin, R. H. , Turk, D. C. , Farrar, J. T. , Haythornthwaite, J. A. , Jensen, M. P. , Katz, N. P. , Kerns, R. D. , Stucki, G. , Allen, R. R. , Bellamy, N. , Carr, D. B. , Chandler, J. , Cowan, P. , Dionne, R. , Galer, B. S. , Hertz, S. , Jadad, A. R. , Kramer, L. D. , Manning, D. C. , … Witter, J. (2005). Core outcome measures for chronic pain clinical trials: IMMPACT recommendations. Pain, 113(1), 9–19. 10.1016/j.pain.2004.09.012 [DOI] [PubMed] [Google Scholar]
  14. Erdek, M. A. , & Pronovost, P. J. (2004). Improving assessment and treatment of pain in the critically ill. International Journal for Quality in Health Care, 16(1), 59–64. 10.1093/intqhc/mzh010 [DOI] [PubMed] [Google Scholar]
  15. Fayaz, A. , Croft, P. , Langford, R. M. , Donaldson, L. J. , & Jones, G. T. (2016). Prevalence of chronic pain in the UK: A systematic review and meta‐analysis of population studies. British Medical Journal Open, 6(6), e010364 10.1136/bmjopen-2015-010364 [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Grunnesjö, M. , Bogefeldt, J. , Blomberg, S. , Delaney, H. , & Svärdsudd, K. (2006). The course of pain drawings during a 10‐week treatment period in patients with acute and sub‐acute low back pain. BMC Musculoskeletal Disorders, 7, 1–9. 10.1186/1471-2474-7-65 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Hadi, M. A. , McHugh, G. A. , & Closs, S. J. (2019). Impact of chronic pain on patients’ quality of life: A comparative mixed‐methods study. Journal of Patient Experience, 6(2), 133–141. 10.1177/2374373518786013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Haskins, B. L. , Lesperance, D. , Gibbons, P. , & Boudreaux, E. D. (2017). A systematic review of smartphone applications for smoking cessation. Translational Behavioral Medicine, 7(2), 292–299. 10.1007/s13142-017-0492-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Hicks, J. L. , Althoff, T. , Sosic, R. , Kuhar, P. , Bostjancic, B. , King, A. C. , Leskovec, J. , & Delp, S. L. (2019). Best practices for analyzing large‐scale health data from wearables and smartphone apps. Npj Digital Medicine, 2(1), 10.1038/s41746-019-0121-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Hüllemann, P. , Keller, T. , Kabelitz, M. , Freynhagen, R. , Tölle, T. , & Baron, R. (2017). Pain drawings improve subgrouping of low back pain patients. Pain Practice, 17(3), 293–304. 10.1111/papr.12470 [DOI] [PubMed] [Google Scholar]
  21. Jamison, R. N. , Fanciullo, G. J. , & Baird, J. C. (2004). Computerized dynamic assessment of pain: Comparison of chronic pain patients and healthy controls. Pain Medicine, 5(2), 168–177. 10.1111/j.1526-4637.2004.04032.x [DOI] [PubMed] [Google Scholar]
  22. Jang, A. , MacLean, D. , & Heer, J. (2014). BodyDiagrams: Improving communication of pain symptoms through drawing. Conference on Human Factors in Computing Systems – Proceedings, 1153–1162. 10.1145/2556288.2557223 [DOI] [Google Scholar]
  23. Krupić, F. , Čustović, S. , Jašarević, M. , Šadić, S. , Fazlić, M. , Grbic, K. , & Samuelsson, K. (2019). Ethnic differences in the perception of pain: A systematic review of qualitative and quantitative research. Medicinski Glasnik, 16(1), 108–114. 10.17392/966-19 [DOI] [PubMed] [Google Scholar]
  24. Lacey, R. J. , Strauss, V. Y. , Rathod, T. , Belcher, J. , Croft, P. R. , Natvig, B. , Wilkie, R. , & McBeth, J. (2015). Clustering of pain and its associations with health in people aged 50 years and older: Cross‐sectional results from the North Staffordshire Osteoarthritis Project. British Medical Journal Open, 5(11), e008389 10.1136/bmjopen-2015-008389 [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Lalloo, C. , Shah, U. , Birnie, K. A. , Davies‐Chalmers, C. , Rivera, J. , Stinson, J. , & Campbell, F. (2017). Commercially available smartphone apps to support postoperative pain self‐management: Scoping review. JMIR MHealth and UHealth, 5(10), e162 10.2196/mhealth.8230 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Leoni, D. , Falla, D. , Heitz, C. , Capra, G. , Clijsen, R. , Egloff, M. , Cescon, C. , Baeyens, J.‐P. , & Barbero, M. (2017). Test‐retest reliability in reporting the pain induced by a pain provocation test: Further validation of a novel approach for pain drawing acquisition and analysis. Pain Practice, 17(2), 176–184. 10.1111/papr.12429 [DOI] [PubMed] [Google Scholar]
  27. Macfarlane, G. J. , Beasley, M. , Smith, B. H. , Jones, G. T. , & Macfarlane, T. V. (2015). Can large surveys conducted on highly selected populations provide valid information on the epidemiology of common health conditions? An analysis of UK Biobank data on musculoskeletal pain. British Journal of Pain, 9(4), 203–212. 10.1177/2049463715569806 [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. McCormick, T. , & Frampton, C. (2019). Assessment of acute and chronic pain. Anaesthesia & Intensive Care Medicine, 20(8), 405–409. 10.1016/j.mpaic.2019.05.010 [DOI] [Google Scholar]
  29. McCurdie, T. , Taneva, S. , Casselman, M. , Yeung, M. , McDaniel, C. , Ho, W. , & Cafazzo, J. (2012). mHealth consumer apps: The case for user‐centered design. Biomedical Instrumentation & Technology, 46(s2), 49–56. 10.2345/0899-8205-46.s2.49 [DOI] [PubMed] [Google Scholar]
  30. Mills, S. E. E. , Nicolson, K. P. , & Smith, B. H. (2019). Chronic pain: A review of its epidemiology and associated factors in population‐based studies. British Journal of Anaesthesia, 123(2), e273–e283. 10.1016/j.bja.2019.03.023 [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Mills, S. , Torrance, N. , & Smith, B. H. (2016). Identification and management of chronic pain in primary care: A review. Current Psychiatry Reports, 18(2), 22 10.1007/s11920-015-0659-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Moher, D. , Liberati, A. , Tetzlaff, J. , & Altman, D. G. (2009). Preferred reporting items for systematic reviews and meta‐analyses: The PRISMA statement. BMJ, 339(7716), 332–336. 10.1136/bmj.b2535 [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Mokkink, L. B. , Terwee, C. B. , Patrick, D. L. , Alonso, J. , Stratford, P. W. , Knol, D. L. , Bouter, L. M. , & de Vet, H. C. W. (2010). The COSMIN study reached international consensus on taxonomy, terminology, and definitions of measurement properties for health‐related patient‐reported outcomes. Journal of Clinical Epidemiology, 63(7), 737–745. 10.1016/j.jclinepi.2010.02.006 [DOI] [PubMed] [Google Scholar]
  34. Najm, A. , Gossec, L. , Weill, C. , Benoist, D. , Berenbaum, F. , & Nikiphorou, E. (2019). Mobile health apps for self‐management of rheumatic and musculoskeletal diseases: Systematic literature review. JMIR MHealth and UHealth, 7(11), e14730 10.2196/14730 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Neubert, T. A. , Dusch, M. , Karst, M. , & Beissner, F. (2018). Designing a tablet‐based software app for mapping bodily symptoms: Usability evaluation and reproducibility analysis. JMIR MHealth and UHealth, 6(5), 10.2196/mhealth.8409 [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. O’Donnell, P. J. , & Curley, H. (1985). Validation of a nonverbal instrument for pain location descriptions in children. Perceptual and Motor Skills, 60(3), 1010 10.2466/pms.1985.60.3.1010 [DOI] [PubMed] [Google Scholar]
  37. Öhlund, C. , Eek, C. , Palmblad, S. , Areskoug, B. , & Nachemson, A. (1996). Quantified pain drawing in subacute low back pain. Spine, 21(9), 1021–1030. 10.1097/00007632-199605010-00005 [DOI] [PubMed] [Google Scholar]
  38. Pogatzki‐Zahn, E. , Schnabel, K. , & Kaiser, U. (2019). Patient‐reported outcome measures for acute and chronic pain. Current Opinion in Anaesthesiology, 32(5), 616–622. 10.1097/ACO.0000000000000780 [DOI] [PubMed] [Google Scholar]
  39. Rowland, S. P. , Fitzgerald, J. E. , Holme, T. , Powell, J. , & McGregor, A. (2020). What is the clinical value of mHealth for patients? Npj Digital Medicine, 3(1), 4 10.1038/s41746-019-0206-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Salazar, A. , de Sola, H. , Failde, I. , & Moral‐Munoz, J. A. (2018). Measuring the Quality of Mobile Apps for the Management of Pain: Systematic Search and Evaluation Using the Mobile App Rating Scale. JMIR MHealth and UHealth, 6(10), e10718 10.2196/10718 [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Schierhout, G. H. , & Myers, J. E. (1996). Is self‐reported pain an appropriate outcome measure in ergonomic‐epidemiologic studies of work‐related musculoskeletal disorders? American Journal of Industrial Medicine, 30(1), 93–98. [DOI] [PubMed] [Google Scholar]
  42. Shaballout, N. , Aloumar, A. , Neubert, T. A. , Dusch, M. , & Beissner, F. (2019). Digital pain drawings can improve doctors’ understanding of acute pain patients: Survey and pain drawing analysis. Journal of Medical Internet Research, 21(1), 1–22. 10.2196/11412 [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Shaballout, N. , Neubert, T.‐A. , Boudreau, S. , & Beissner, F. (2019). From paper to digital applications of the pain drawing: Systematic review of methodological milestones. JMIR MHealth and UHealth, 7(9), e14569 10.2196/14569 [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Smith, D. , Wilkie, R. , Croft, P. , & McBeth, J. (2018). Pain and mortality in older adults: The influence of pain phenotype. Arthritis Care & Research, 70(2), 236–243. 10.1002/acr.23268 [DOI] [PubMed] [Google Scholar]
  45. Southerst, D. , Côté, P. , Stupar, M. , Stern, P. , & Mior, S. (2013). The reliability of body pain diagrams in the quantitative measurement of pain distribution and location in patients with musculoskeletal pain: A systematic review. Journal of Manipulative and Physiological Therapeutics, 36(7), 450–459. 10.1016/j.jmpt.2013.05.021 [DOI] [PubMed] [Google Scholar]
  46. Southerst, D. , Stupar, M. , Côté, P. , Mior, S. , & Stern, P. (2013). The reliability of measuring pain distribution and location using body pain diagrams in patients with acute whiplash‐associated disorders. Journal of Manipulative and Physiological Therapeutics, 36(7), 395–402. 10.1016/j.jmpt.2013.05.023 [DOI] [PubMed] [Google Scholar]
  47. Stinson, J. N. , Lalloo, C. , Harris, L. , Isaac, L. , Campbell, F. , Brown, S. , Ruskin, D. , Gordon, A. , Galonski, M. , Pink, L. R. , Buckley, N. , Henry, J. L. , White, M. , & Karim, A. (2014). iCanCope with painTM: User‐centred design of a web‐ and mobile‐based self‐management program for youth with chronic pain based on identified health care needs. Pain Research and Management, 19(5), 257–265. 10.1155/2014/935278 [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Stoyanov, S. R. , Hides, L. , Kavanagh, D. J. , Zelenko, O. , Tjondronegoro, D. , & Mani, M. (2015). Mobile app rating scale: A new tool for assessing the quality of health mobile apps. JMIR MHealth and UHealth, 3(1), e27 10.2196/mhealth.3422 [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Tsang, A. , Von Korff, M. , Lee, S. , Alonso, J. , Karam, E. , Angermeyer, M. C. , Borges, G. L. G. , Bromet, E. J. , de Girolamo, G. , de Graaf, R. , Gureje, O. , Lepine, J.‐P. , Haro, J. M. , Levinson, D. , Oakley Browne, M. A. , Posada‐Villa, J. , Seedat, S. , & Watanabe, M. (2008). Common chronic pain conditions in developed and developing countries: Gender and age differences and comorbidity with depression‐anxiety disorders. The Journal of Pain, 9(10), 883–891. 10.1016/j.jpain.2008.05.005 [DOI] [PubMed] [Google Scholar]
  50. Vo, V. , Auroy, L. , & Sarradon‐Eck, A. (2019). Patients’ perceptions of mHealth apps: Meta‐ethnographic review of qualitative studies. JMIR MHealth and UHealth, 7(7), e13817 10.2196/13817 [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. von Baeyer, C. L. , Lin, V. , Seidman, L. C. , Tsao, J. C. , & Zeltzer, L. K. (2011). Pain charts (body maps or manikins) in assessment of the location of pediatric pain. Pain Management, 1(1), 61–68. 10.2217/pmt.10.2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Wenngren, A. , & Stålnacke, B.‐M. (2009). Computerized assessment of pain drawing area: A pilot study. Neuropsychiatric Disease and Treatment, 5, 451–456. http://www.ncbi.nlm.nih.gov/pubmed/19721724 [PMC free article] [PubMed] [Google Scholar]
  53. Wood, S. (2004). Factors influencing the selection of appropriate pain assessment tools. Nursing times, 100(35), 42–47. http://www.ncbi.nlm.nih.gov/pubmed/15471273 [PubMed] [Google Scholar]
  54. Vos, T. , Flaxman, A. D. , Naghavi, M. , Lozano, R. , Michaud, C. , Ezzati, M. , Shibuya, K. , Salomon, J. A. , Abdalla, S. , Aboyans, V. , Abraham, J. , Ackerman, I. , Aggarwal, R. , Ahn, S. Y. , Ali, M. K. , Alvarado, M. , Anderson, H. R. , Anderson, L. M. , Andrews, K. G. , … Murray, C. J. (2012). Years lived with disability (YLDs) for 1160 sequelae of 289 diseases and injuries 1990–2010: A systematic analysis for the global burden of disease study 2010. The Lancet, 380(9869), 2163–2196. 10.1016/S0140-6736(12)61729-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Younger, J. , McCue, R. , & Mackey, S. (2009). Pain outcomes: A brief review of instruments and techniques. Current Pain and Headache Reports, 13(1), 39–43. 10.1007/s11916-009-0009-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Zhao, P. , Yoo, I. , Lancey, R. , & Varghese, E. (2019). Mobile applications for pain management: An app analysis for clinical usage. BMC Medical Informatics and Decision Making, 19(1), 1–10. 10.1186/s12911-019-0827-7 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Table S1‐S3


Articles from European Journal of Pain (London, England) are provided here courtesy of Wiley

RESOURCES