Skip to main content
. 2019 Apr 1;5(1):e10187. doi: 10.2196/10187

Table 2.

Usability testing of the app (N=175 clinicians who used each resource).

Type of testing n (%)
Usability testing of the BTcPa diagnostic workflow

Definitions of BTcP by Davies et al [6]/Escobar et al [16] 146 (83.4)

Davies algorithm 141 (80.7)

Other considerations of BTcP diagnosis 46 (26.4)

DN4 neuropathic scaleb 122 (69.5)

Reviewed Edmonton scaleb 111 (63.4)
Usability testing of the app’s pain tools

Opioid rotation 92 (52.8)

Karnofsky scalec 142 (81.3)

ECOGd scale of performance statusc 147 (84.1)

Visual Numeric Scalee 163 (93.3)

Categorical Scalee 124 (70.8)

Charlson Comorbidity Index 97 (55.4)

PQRSTf questionnaire 111 (63.7)
Usability testing of App INES·DIO

Would you include the app report with the clinical history of the patient?


Always 65 (37.1)


Almost always 54 (30.9)


Occasionally 52 (29.7)


Never 4 (2.3)

Would you use App INES·DIO in a different patient profile?


Most likely 70 (40.0)


Likely 52 (29.7)


Least likely 49 (28.0)


Unlikely 4 (2.3)

How many days have you used App INES·DIO on average?


Everyday 31 (17.7)


4-6 times per week 38 (21.7)


2-3 times per week 74 (42.3)


Once per week 32 (18.3)

aBTcP: breakthrough cancer pain.

bTools to assess baseline pain.

cFunctional scales.

dECOG: Eastern Cooperative Oncology Group.

ePain scales.

fPQRST: (P) provocative and palliative factors; (Q) qualitative description of pain; (R) region and radiation of pain; (S) severity or intensity of pain after being scored by means of VNS and CS; (T) timing or pain changes over time.