Skip to main content
Journal of Medical Internet Research logoLink to Journal of Medical Internet Research
. 2025 Jan 20;27:e53503. doi: 10.2196/53503

A New Smartphone-Based Cognitive Screening Battery for Multiple Sclerosis (icognition): Validation Study

Stijn Denissen 1,2,✉,#, Delphine Van Laethem 1,3,#, Johan Baijot 1, Lars Costers 1,2, Annabel Descamps 2, Ann Van Remoortel 1,4, Annick Van Merhaegen-Wieleman 1,5, Marie D’hooghe 4,6, Miguel D'Haeseleer 4,5,6, Dirk Smeets 1,2, Diana M Sima 1,2, Jeroen Van Schependom 1,7, Guy Nagels 1,5,8
Editor: Amaryllis Mavragani
Reviewed by: Mt Ghozali, Omid Mirmosayyeb, Felipe Balistieri Santinelli
PMCID: PMC11791456  PMID: 39832354

Abstract

Background

Cognitive deterioration is common in multiple sclerosis (MS) and requires regular follow-up. Currently, cognitive status is measured in clinical practice using paper-and-pencil tests, which are both time-consuming and costly. Remote monitoring of cognitive status could offer a solution because previous studies on telemedicine tools have proved its feasibility and acceptance among people with MS. However, existing smartphone-based apps include designs that are prone to motor interference and focus primarily on information processing speed, although memory is also commonly affected.

Objective

This study aims to validate a smartphone-based cognitive screening battery, icognition, to detect deterioration in both memory and information processing speed.

Methods

The icognition screening battery consists of 3 tests: the Symbol Test for information processing speed, the Dot Test for visuospatial short-term memory and learning, and the visual Backward Digit Span (vBDS) for working memory. These tests are based on validated paper-and-pencil tests: the Symbol Digit Modalities Test, the 10/36 Spatial Recall Test, and the auditory Backward Digit Span, respectively. To establish the validity of icognition, 101 people with MS and 82 healthy participants completed all tests. Of the 82 healthy participants, 20 (24%) repeated testing 2 to 3 weeks later. For each icognition test, validity was established by the correlation with its paper-and-pencil equivalent (concurrent validity), the correlation and intraclass correlation coefficient (ICC) between baseline and follow-up testing (test-retest reliability), the difference between people with MS and healthy participants, and the correlation with other clinical parameters such as the Expanded Disability Status Scale.

Results

All icognition tests correlated well with their paper-and-pencil equivalents (Symbol Test: r=0.67; P<.001; Dot Test: r=0.31; P=.002; vBDS: r=0.69; P<.001), negatively correlated with the Expanded Disability Status Scale (Symbol Test: ρ=–0.34; P<.001; Dot Test: ρ=−0.32; P=.003; vBDS: ρ=−0.21; P=.04), and showed moderate test-retest reliability (Symbol Test: ICC=0.74; r=0.85; P<.001; Dot Test: ICC=0.71; r=0.74; P<.001; vBDS: ICC=0.72; r=0.83; P<.001). Test performance was comparable between people with MS and healthy participants for all cognitive tests, both in icognition (Symbol Test: U=4431; P=.42; Dot Test: U=3516; P=.32; vBDS: U=3708; P=.27) and the gold standard paper-and-pencil tests (Symbol Digit Modalities Test: U=4060.5, P=.82; 10/36 Spatial Recall Test: U=3934; P=.74; auditory Backward Digit Span: U=3824.5, P=.37).

Conclusions

icognition is a valid tool to remotely screen cognitive performance in people with MS. It is planned to be included in a digital health platform that includes volumetric brain analysis and patient-reported outcome measures. Future research should establish the usability and psychometric properties of icognition in a remote setting.

Keywords: multiple sclerosis, telemedicine, cognition, memory, information processing speed, mobile phone

Introduction

Background

Medicine is increasingly digitalizing, and there are compelling reasons to stimulate this trend. Clinicians can more easily access and share electronic health records, and storing data in a digital format facilitates visualization and organization in research databases, yielding new insights into pathology and disease management. Moreover, electronic health records drive artificial intelligence research [1], while artificial intelligence, in turn, further stimulates storing records digitally [2], closing a positive feedback loop. Far from being a mere “nice to have,” digital medicine was crucial during the COVID-19 pandemic [3], enabling telemedicine services to be provided when social distancing was essential.

Telemedicine provides practical solutions for people with multiple sclerosis (MS), a chronic disease characterized by inflammation and degeneration of the central nervous system [4]. Telemedicine tools are well accepted by patients, and their feasibility and cost-effectiveness have been established previously [5]. Moreover, patients tend to objectively benefit from the use of these tools, which can, for example, aid in fatigue management [6] and improve cognitive function [7]. The latter is important because nearly half of the people with MS have cognitive impairment [8], which has significant repercussions on daily life activities, societal participation, employment, and susceptibility to psychiatric disorders [9].

Telemedicine could also facilitate cognitive monitoring of people with MS. First, it would allow increasing temporal resolution of cognitive trajectories because cognitive tests are usually performed during routine follow-up sessions that are months apart. This is problematic because cognitive decline can be sudden, unexpected, and severe (eg, in case of a relapse) [10,11]. Remote testing could help detect minimal changes early, enabling timely intervention. Second, it would unburden clinicians because current practice relies on paper-and-pencil tests administered under the supervision of a trained examiner.

Smartphones especially provide a window of opportunity, with an estimated 6.7 billion subscriptions worldwide (69% of the population) [12]. However, current smartphone-based cognitive assessments focus primarily on information processing speed (IPS) [13]; yet, besides slowed IPS, the hallmark cognitive problem in MS is impaired memory [14], for which the first smartphone test was only recently introduced by Podda et al [15]. Memory assessments emerged earlier on tablet devices [16] but are less suitable for consistent follow-up because tablet devices are used far less frequently than smartphones. Furthermore, smartphone tests might be prone to motor interference. They are predominantly digital versions of the Symbol Digit Modalities Test (SDMT) [17], which is a popular test in clinical practice to measure IPS and has excellent psychometric properties [18]. Digitalizing the SDMT allows randomizing its key, which could reduce practice effect as reported in the study by Pereira et al [19]. However, creating an exact digital replica of the SDMT requires patients to choose from 9 small buttons on the screen, which could cause motor interference because fine motor skills are commonly affected in MS [20].

Objectives

To tackle these limitations, in this study, we aim to validate a new smartphone-based cognitive screening battery called “icognition.” It is a quick, smartphone-based screening tool for remote follow-up of the 2 most commonly impaired cognitive domains in MS: IPS and memory [14]. It is intended to be part of the recently established icompanion app, a digital diary for people with MS [21]. Regular remote screening could enable faster confirmation of cognitive deterioration by a neuropsychologist and prompt intervention by the patient’s neurologist.

Methods

Study Design

This is an observational case-control study designed to validate the icognition smartphone app.

Participants

Study participants were recruited between June 17, 2021, and January 3, 2023. People with MS were enrolled from the outpatient clinics of the neurology department at Universitair Ziekenhuis Brussel (secondary care) and the National MS Center at Melsbroek (tertiary care). They were recruited by the local study nurse during their follow-up visit. Healthy control participants were recruited via leaflets and the social networks of the researchers involved. Inclusion criteria for people with MS were a confirmed diagnosis of MS according to the McDonald criteria [22]. People with MS were excluded if they had been hospitalized for reasons other than rehabilitation or if they had experienced a relapse within the past month. Both people with MS and healthy control participants were excluded if they had any other neurological or psychiatric disorder or learning disorder. A total of 101 people with MS and 82 healthy control participants (matched on age, sex, and education level) met the inclusion and exclusion criteria for this study. All participants were either Dutch-speaking or bilingual, including Dutch, and were aged ≥18 years.

Ethical Considerations

This study was approved by the medical ethics committee of Universitair Ziekenhuis Brussel (BUN 143201940335) and the National MS Center at Melsbroek. All participants signed informed consent (in Dutch) before inclusion. Data were pseudonymized and stored in the protected OneDrive cloud service (Microsoft Corporation) of Vrije Universiteit Brussel. When publicly shared via GitHub, as explained in the Data Availability section, the data will be anonymized. This manuscript presents the primary analysis of these data. Participants did not receive compensation for taking part in the study.

icognition Screening Battery

The icognition cognitive screening battery consists of 3 tests (Figure 1).

Figure 1.

Figure 1

Screenshots of the icognition tests (although the instructions were in Dutch during testing, they are presented here in English). (A) The Symbol Test for information processing speed (score=number of correct responses [does the symbol combination occur in the key?] in 90 s). (B) The Dot Test for visuospatial short-term learning and memory (score=total number of correctly indicated dots across 10 trials [maximum score=30]). (C) Visual Backward Digit Span (vBDS) for working memory (score=sum of span lengths of correct spans [correct inversion of shown span]).

The Symbol Test is based on the computerized Digit Symbol Substitution Test presented in the study by Rypma et al [23]. In the Symbol Test, a combination of symbols is presented to the participants, one at a time. A key, consisting of 9 pairs of symbols, is displayed at the top and is shuffled for each trial. For each trial, the participant needs to indicate whether the presented combination appears in the key. The total score is the number of correct answers provided in 90 seconds. This test is designed to assess IPS.

The Dot Test consists of 3 phases. In the first phase, a participant is presented a 4×4 grid in which 3 dots are shown for 3 seconds. Next, as a distractor task, the participant is shown a 4×6 grid of “E” and “F” shapes and must identify as many “F” shapes as possible in 4 seconds. In the last phase, the participant must indicate in an empty 4×4 grid where the 3 dots of the first phase were located. The Dot Test is inspired by the Dot Memory Test presented in the study by Sliwinski et al [24], with all grids reduced in size compared to their version (5×5 grids). We also implemented a criterion for the distractor task, requiring the participant to identify at least 3 “F” shapes. If this criterion was not met, the trial was restarted. The 3 dots could not be aligned on 1 line or form an L-shape within a 2×2 block of cells. The total score is the number of correctly indicated dots across 10 trials. This test is designed to assess visuospatial short-term memory and learning.

In the visual Backward Digit Span (vBDS), a series of digits is presented on the screen one by one, each for 1 second, as described in the study by Hilbert et al [25]. The participant must then list the digits in reverse order. Spans were randomly generated with digits between 0 and 9, with the following constraints: a digit can appear only once in the span, and a chain of ≥3 digits cannot have a fixed increment or decrement of 1 or 2 digits, in accordance with Woods et al [26]. Scoring is based on the sum of all correct span lengths; for example, if a participant correctly recalls 2 spans of length 3 and 1 span of length 4, the total score is 10. This test is designed to assess working memory.

All tests were performed on a Samsung Galaxy A10 smartphone (6.2-inch screen size) and were supervised by a test examiner. Each test was directly preceded by a practice phase to familiarize the participants with the respective test. This phase consisted of 5 trials for the Symbol Test, 3 for the Dot Test, and 4 (2 of length 3 and 2 of length 4) for the vBDS. The Symbol Test is performed with the smartphone in landscape orientation, whereas for the Dot Test and vBDS, the smartphone must be in portrait orientation. In the design of icognition, careful consideration was given to the potential biasing influence of fine motor impairment in MS [20]. Motor interference was minimized by using 2 large buttons for the Symbol Test (in contrast to digital SDMT variants where participants must choose from 9 smaller buttons [27]) and not placing any restrictions on the response time in the other icognition tests.

Validation Procedure

The procedure to validate icognition is based on the study by Benedict et al [28] and involves assessing 4 criteria, as outlined in the following subsections.

Concurrent Validity

We assessed how well each icognition test correlates with its paper-and-pencil equivalent. For the Symbol Test, the equivalent was the SDMT [17]. In the SDMT, a sheet is presented to the participant with a key of 9 symbol-digit pairs at the top and a list of symbols without corresponding digits. In 90 seconds, the participant must convert as many symbols to digits as possible from the list, reading them out loud to the examiner, using the key. The Dot Test is based on the 10/36 Spatial Recall Test (SPART) [29]. In the SPART, the participant is shown a 6×6 grid with a pattern of 10 dots for 10 seconds. Subsequently, the grid is removed, and the participant is asked to replicate the pattern using 10 checkers. This process is repeated 3 times. The final score is the total number of correctly placed checkers across all trials. Finally, for the vBDS, a modified version of the Wechsler Adult Intelligence Scale, Fourth Edition, auditory Backward Digit Span [30] was used. In the auditory Backward Digit Span, digit spans are read out loud to the participant, who is asked to repeat them in reverse order. The original test consists of 2 trials for each span length, starting with a span of 2 digits and increasing by 1 digit each time the participant correctly completes at least 1 of the 2 trials. As discussed in Woods et al [31], in the original Wechsler Adult Intelligence Scale, Fourth Edition, design, participants with the same score can have a different number of correct spans. To mitigate this, we used a fixed number of spans, ranging from 3 to 7 digits in length (Table S1 in Multimedia Appendix 1). The complete list was always administered for each participant. The scoring metric is the same as described earlier for the icognition Backward Digit Span.

Test-Retest Reliability

Benedict et al [28] mention that test-retest reliability should be assessed on a “small sample” of either people with MS or healthy controls [28]. We aimed to retest the healthy controls 2 to 3 weeks after baseline testing. We used the intraclass correlation coefficient (ICC) to assess agreement between baseline and retesting, which is explained in more detail in the Statistical Analyses subsection.

Comparison of Performance

For each test, we compared the performance of people with MS to that of age-, sex-, and education level–matched healthy controls using the Mann-Whitney U test. The analysis was then repeated after correcting test performance for age, sex, and education level (Figure S1 in Multimedia Appendix 1). Details of the correction methodology are presented in the Statistical Analyses subsection.

Assessment of Correlations

Finally, we assessed the correlation of each icognition test with the Expanded Disability Status Scale (EDSS) [32], disease duration, the Beck Depression Inventory (BDI) [33], the Fatigue Scale for Motor and Cognitive Functions (FSMC) [34], education level, and age.

Data Curation

Data were entered independently by 2 researchers (SD and DVL). Conflicts in data entry were resolved through mutual discussion.

Statistical Analyses

We used an α level of .05 for all analyses. Participants with missing data on a certain test were only excluded for that specific test. We used the Spearman correlation for nonlinear and categorical variables (EDSS, BDI, FSMC, and education level), whereas the Pearson correlation was used otherwise. Mann-Whitney U tests were used for between-group distribution comparisons.

To assess test-retest reliability, following the approach of van Oirschot et al [35] and the guidelines by Koo and Li [36], we used the following ICC type: “two-way mixed effects, absolute agreement, single rater/measurement” (ICC[A,1]). In this study, the smartphone app served as the sole “rater,” which is why we used a 2-way mixed effects model, appropriate when the “selected raters are the only raters of interest” [36]. For the same reason, the “type” of ICC was “single rater/measurement,” while for “definition,” we used “absolute agreement,” reflecting the extent to which 1 rater’s score (in our case, baseline testing) equals the other rater’s score (in our case, retesting). The decision process is illustrated in the decision flowchart in the study by Koo and Li [36]. To interpret the magnitude of the test-retest reliability, we used the guidelines by Koo and Li [36]: poor (<0.50), moderate (0.50-0.75), good (0.75-0.90), and excellent (>0.90).

The magnitude of correlations was interpreted using the classification by Portney and Watkins [37]: small (<0.25), fair (0.25-0.50), moderate to good (0.50-0.75), and excellent (>0.75).

Finally, for each cognitive test, we fitted a regression equation with age, sex, and education level as independent variables and test performance as dependent variable on the healthy control data. This allows calculation of the expected score of a participant, given their age, sex, and education level, and subsequently comparison with the actual score of the participant, resulting in a z score. The technical details of the procedure, including the required values for performing z normalization on other data, are presented in Multimedia Appendix 1, while Table S2 in Multimedia Appendix 1 presents the values necessary to perform the procedure. Test performance was considered impaired if the z score was ≤−1.5, based on the study by Benedict et al [38].

Results

Overview

Participant characteristics are presented in Table 1.

Table 1.

Participant characteristics (n=183)a.


People with MSb (n=101) Healthy controls (n=82) P value
Demographics

Age (y), mean (SD) 45.4 (10.0) 46.8 (14.7) .59c

Sex, n (%) .74d


Female 63 (62.4) 54 (65.9)


Male 38 (37.6) 28 (34.1)

Education, median (IQR) 15 (12-17) 15 (13-17) .74c
MS specific

Disease duration (y), mean (SD) 11.7 (7.4) e

MS type, n (%)


RRMSf 86 (85.1)



SPMSg 8 (7.9)



PPMSh 7 (6.9)


EDSSi , median (IQR) 3 (2-4)
Paper-and-pencil tests

SDMTj , mean (SD) 58.5 (10.0) 58.3 (9.9) .82c


Those considered impaired, n (%) 7 (6.9) 6 (7.3) .99d

10/36 SPARTk , mean (SD) 20.6 (4.3) 20.1 (4.6) .74c


Those considered impaired, n (%) 8 (7.9) 4 (4.9) .60d

Auditory BDSl , mean (SD) 48.4 (18.7) 46.0 (17.6) .37c


Those considered impaired, n (%) 9 (8.9) 6 (7.3) .90d

BDIm , mean (SD) 9.3 (5.7) 5.8 (5.1) <.001c

FSMCn , mean (SD) 58.7 (16.4) 39.9 (12.1) <.001c
icognition tests

Symbol Test, mean (SD) 24.8 (6.3) 25.4 (6.4) .42c


Number impaired, n (%) 13 (12.9) 7 (8.5) .49d

Dot Test, mean (SD) 21.8 (5.1) 21.2 (4.9) .32c


Number impaired, n (%) 9 (8.9) 7 (8.5) .99d

Visual BDS, mean (SD) 46.9 (16.9) 43.8 (16.7) .27c


Number impaired, n (%) 8 (7.9) 9 (11) .65d

aThe following variables had missing values: Expanded Disability Status Scale (n=7), 10/36 Spatial Recall Test (people with multiple sclerosis: n=1; healthy controls: n=1), Beck Depression Inventory (people with multiple sclerosis: n=2; healthy controls: n=1), Fatigue Scale for Motor and Cognitive Functions (people with multiple sclerosis: n=2; healthy controls: n=7), Dot Test (people with multiple sclerosis: n=3; healthy controls: n=1), and visual Backward Digit Span (people with multiple sclerosis: n=1). An additional 3 participants were excluded from the Dot Test analyses because they had been tested with an earlier version of icognition where a larger grid size was used. We reduced the grid size after these 3 participants were tested because we deemed this test to be too difficult. However, these 3 participants were included when the mean (SD) and number of participants considered impaired was calculated.

bMS: multiple sclerosis.

cMann-Whitney U test.

dChi-square test.

eNot applicable.

fRRMS: relapsing-remitting multiple sclerosis.

gSPMS: secondary progressive multiple sclerosis.

hPPMS: primary progressive multiple sclerosis.

iEDSS: Expanded Disability Status Scale

jSDMT: Symbol Digit Modalities Test.

kSPART: 10/36 Spatial Recall Test.

lBDS: Backward Digit Span

mBDI: Beck Depression Inventory.

nFSMC: Fatigue Scale for Motor and Cognitive Functions.

Concurrent Validity

Figure 2 shows the scatterplot of each icognition test with its paper-and-pencil equivalent. The Symbol Test showed a significant moderate to good correlation with SDMT performance (healthy controls: r=0.68; P<.001; people with MS: r=0.67; P<.001). There was also a significant fair correlation between the Dot Test and the SPART (healthy controls: r=0.30; P=.007; people with MS: r=0.31; P=.002) and a moderate to good correlation between the vBDS and its auditory equivalent (healthy controls: r=0.69; P<.001; people with MS: r=0.69; P<.001).

Figure 2.

Figure 2

Concurrent validity. Scatterplots comparing the scores of each icognition test (x-axis) with those of its corresponding paper-and-pencil equivalent (y-axis). Correlations for both people with multiple sclerosis (MS) and healthy controls were (A) moderate to good, (B) fair, and (C) moderate to good. aBDS: auditory backward digit span; SDMT: Symbol Digit Modalities Test; SPART: 10/36 Spatial Recall Test; vBDS: visual Backward Digit Span.

Test-Retest Reliability

In total, 20 healthy controls were retested with an average intertest interval of 18 (SD 3, range 14-23) days after initial testing to establish test-retest reliability. For the Dot Test, there was 1 missing value at baseline testing (n=19). Test-retest reliability (Figure 3) was moderate for the Symbol Test (ICC=0.74, r=0.85; P<.001), moderate for the Dot Test (ICC=0.71, r=0.74; P<.001), and moderate for the vBDS (ICC=0.72, r=0.83; P<.001).

Figure 3.

Figure 3

Test-retest reliability. Scatterplots comparing the scores of each icognition test at baseline (x-axis) with the retest scores an average of 18 (SD 3) days later (y-axis). All tests demonstrated moderate test-retest reliability. vBDS: visual Backward Digit Span.

Difference Between People With MS and Healthy Controls

For all icognition tests, there was no significant difference in performance between healthy controls and people with MS (Symbol Test: U=4431; P=.42; Dot Test: U=3516; P=.32; vBDS: U=3708; P=.27; Figure 4). The results were similar for all paper-and-pencil tests (SDMT: U=4060.5; P=.82; SPART: U=3934; P=.74; auditory Backward Digit Span: U=3824.5; P=.37; Figure S2 in Multimedia Appendix 1).

Figure 4.

Figure 4

Comparison of the performance of healthy controls and people with multiple sclerosis (MS) on the icognition tests. Test performance was not significantly different between people with MS and healthy controls for all icognition tests. vBDS: visual Backward Digit Span.

Correlations With Clinical Parameters

A correlation matrix between the icognition tests and different clinical tests is presented in Table 2. In general, higher scores on the icognition tests were associated with less physical disability (measured using the EDSS), younger age, and a higher education level. Furthermore, all tests except the Symbol Test showed a correlation with disease duration, whereas fatigue (measured using the FSMC) was negatively associated with Dot Test performance. No correlation was found between any icognition test and depression (measured using the BDI). Correlation magnitudes were small or fair, with the exception of the moderate to good correlation between the Symbol Test and age (|r|=0.52).

Table 2.

Correlation matrix of icognition tests with clinical variables. The Spearman correlation was used for the Beck Depression Inventory (BDI), Fatigue Scale for Motor and Cognitive Functions (FSMC), Expanded Disability Status Scale (EDSS), and education level, while the Pearson correlation was used for age and disease duration.


Symbol Test Dot Test vBDSa

Correlation P value Correlation P value Correlation P value
Age −0.52 <.001 −0.29 .004 −0.27 .008
Education level 0.20 .04 0.37 <.001 0.22 .03
BDI −0.03 .78 −0.12 .27 0.03 .80
FSMC −0.10 .31 −0.26 .01 −0.03 .79
EDSS −0.34 <.001 −0.32 .003 −0.21 .04
Disease duration −0.17 .08 −0.28 .006 −0.31 .002

avBDS: visual Backward Digit Span.

Discussion

Overview

In this paper, we present the results of the validation process of a smartphone-based screening battery for cognitive problems in people with MS. All tests correlated with their paper-and-pencil equivalents (concurrent validity), although correlation between the Dot Test and the SPART was only fair, most likely due to the fact that the SPART—the gold standard—lacks a distractor phase. The distractor phase was initially added to avoid ceiling effects and increase sensitivity in detecting cognitive decline. However, other test differences between the Dot Test and the SPART, such as grid size and the number of trials, could also cause the correlation to be weaker. Furthermore, all tests correlated with various clinical variables (age, education level, and EDSS) and showed moderate test-retest reliability. These findings indicate the suitability of the battery for routine remote screening of cognitive problems, a condition that many people with MS are likely to develop over time [8].

State of the Art

We performed a systematic search (Multimedia Appendix 1) to identify smartphone-based cognitive tests for MS, which are summarized in Table 3. In brief, these apps began to emerge in 2020 and predominantly use a typical SDMT design. In this format, the mobile phone is held in landscape orientation, and a key of 9 symbol-digit pairs is displayed at the top. For 90 seconds, 1 symbol at a time is shown in the center of the screen, and the participant selects as quickly as possible the matching digit from 9 digit buttons located at the bottom of the screen. These tests are primarily designed to measure IPS, although they are also used to measure cognitive fatigability [39,40]. Other tests designed to measure specific cognitive domains include the smartphone version of the Trail Making Test Part B [41] for executive function and the Auditory Test of Processing Speed for IPS [42]. Furthermore, smartphone keystroke analysis [43,44] and various cognitive training games [45] were not designed for a specific cognitive domain, but their relationships with multiple cognitive domains was assessed. Recently, the first smartphone-based cognitive screening battery (DIGICOG-MS [15]) was introduced, marking an important next step in smartphone-based cognitive assessment.

Table 3.

Summary of the state of the art in smartphone-based cognitive tests for people with multiple sclerosis (MS). The “Target domain” column indicates the specific cognitive domain assessed by each cognitive test.

Name Summary Target domain
MCTa [46] The MCT, part of MSCopilot, is a digital version of the SDMTb. In the MCT, a key displaying 9 symbol-digit pairs is presented at the top of the screen, with an exit button located in the top right corner. One symbol at a time is presented below the key. No details pertaining to the length of the test or scoring method are provided in the paper. Information processing speed
sSDMTc [35] The sSDMT, part of MS Sherpa, is a digital version of the SDMT. In the sSDMT, a symbol-digit key (with 9 pairs) is presented at the top of the screen. A single symbol is then presented in the center of the screen, and participants must quickly tap the correct matching digit from 9 digit buttons presented at the bottom. The total score is the number of correct digits selected in 90 s. A timer is displayed, and the key is randomized at each trial. Information processing speed
Voice-controlled DSSTd [47] The voice-controlled DSST is part of the elevateMS app. Although limited details are provided, the app seems to be a variant of the SDMT, in which responses are collected via microphone. Information processing speed and working memory
sSDMT [27] The sSDMT, part of the Neurological Functional Test Suite, resembles the sSDMT (described previously). However, 3 differences are noteworthy. First, no timer is provided. Second, the test duration is 75 s; a 90-s performance score is calculated by multiplying the result by 90/75. Finally, the app also supports vocalized responses using a microphone and voice recognition to test participants with severe motor impairments. Information processing speed
Mobile Trail Making Test Part B [41] The app displays 13 circles, 6 containing a letter and 7 containing a number. The goal of the test is to connect the circles as quickly as possible in order, alternating between numbers and letters (1-A-2-B and so on, up to 7). The smartphone test was introduced by Ross et al [48] and assessed in people with MS in the study by Chen et al [41]. Executive function
Neurokeys [43] Neurokeys is an alternative keyboard that records typing events (press and release) and subsequently calculates a range of features such as press duration and time interval between 2 presses. No specific cognitive domain
BiAffect [44] Similar to Neurokeys, the BiAffect app features a custom keyboard that temporarily replaces the smartphone’s default keyboard and is used to collect keystrokes and subsequently calculate various features. No specific cognitive domain
cFASTe [40] The cFAST is similar to the sSDMT but includes an exit button and a progress bar to indicate the remaining response time. Unlike the 90-s duration of the sSDMT, the cFAST test lasts 5 minutes because the goal is to assess cognitive fatigability rather than information processing speed. Cognitive fatigability
eSDMTf [49] The eSDMT, part of Floodlight, is an electronic version of the SDMT. Although few details are provided, the app seems to use a similar design to the sSDMT [39,50]. However, no timer or randomization of the key is mentioned. Information processing speed
Cognitive games in the dreaMS app [45] Several cognitive games were included in the dreaMS app featured in the study by Pless et al [45]. No specific cognitive domain
Auditory Test of Processing Speed [42] Participants are presented auditory digits between 1 and 99 across 3 trials (20 digits each), with each trial involving different questions for the participants to answer (ie, trial 1: is the digit greater than 50? Trial 2: is the digit greater than 50 or odd? Trial 3: is the digit greater than 50 and odd?). The test examiner records the participants’ responses, and response time is calibrated before testing. Information processing speed
MS Care Connect [51] The MS Care Connect app is freely available for remote health assessments, including cognitive tests, in people with MS. Although limited information is available about the cognitive tests, a scoping review by Michaud et al [52] mentions it contains an SDMT-like test. According to a video on the app’s website, the test is administered in portrait orientation, with the key (9 symbol-digit combinations) at the top of the screen, a symbol presentation field below it, and a 3×3 answer keypad at the bottom [53]. Information processing speed
Konectom [54] The test is designed to measure cognitive processing speed, with the smartphone held in portrait orientation. A symbol-digit key (with 9 pairs) is displayed at the top of the screen and remains fixed for the 90-s test duration. Symbols are presented in the center of the screen, and a 3×3 answer keyboard (digits 1-9) is located at the bottom. To correct for visuomotor interference, a separate test is included in which participants are asked to tap, in the same 3×3 answer keyboard and for 20 s, as many numbers as possible that are consecutively displayed in the center of the screen. Information processing speed
DIGICOG-MS [15] This is a smartphone-based cognitive battery with tests for memory, semantic fluency, and information processing speed. All tests seem to be designed for use in portrait orientation. A symbol-digit key (with 9 pairs) is presented at the top of the screen and remains fixed for the 90-s test duration. Four symbols are shown at a time in the center of the screen, and a 3×3 answer keyboard (digits 1-9) is located at the bottom. Visuospatial memory, verbal memory, semantic fluency, and information processing speed

aMCT: Mobile Cognition Test.

bSDMT: Symbol Digit Modalities Test.

csSDMT: smartphone Symbol Digit Modalities Test.

dDSST: Digit Symbol Substitution Test.

ecFAST: Cognitive Fatigability Assessment Test.

feSDMT: electronic Symbol Digit Modalities Test.

The icognition screening battery differs from the state of the art in 2 ways. First, fine motor and visual problems are frequently reported in people with MS [55,56]. These symptoms are likely to interfere with time-based digital processing speed tests and could be affected by design considerations, as seen in the common digital SDMT design. This design includes multiple answer buttons (typically 9), and if configured in portrait orientation may result in a small symbol-digit key, especially on smaller devices. Although the problem is recognized by Woelfle et al [50] and the Scaramozza et al [54], the latter implementing a separate test to quantify motor interference, we opted to minimize motor and visual interference through design. Specifically, we used a simplistic design in landscape orientation, featuring 2 large answer buttons (Figure 1). Second, rather than a single test, icognition features a battery of tests designed to capture a broader cognitive profile because cognitive problems in people with MS are not limited to slowed IPS [14]. To the best of our knowledge, only Podda et al [15] have recently published a smartphone-based cognitive battery for people with MS. Moreover, icognition will be included in a digital care platform assessing symptoms, imaging biomarkers, and patient-reported outcomes [21] to provide a holistic picture of a patient’s well-being.

A Cognitively Preserved Sample of People With MS

People with MS scored equally well on all tests compared to healthy controls. However, as can be observed in Table 1 and Figure S2 in Multimedia Appendix 1, this was also the case for the paper-and-pencil tests. Indeed, we seem to have included a sample of people with MS with relatively preserved cognition. We tested this in a post hoc analysis using the ANOVA method described by Anders [57], comparing the SDMT performance scores of our sample of people with MS (mean 58.5, SD 10.0; n=101; Table 1) with those reported in previous studies. López-Góngora et al [58] report an average SDMT performance score of 54.3 (SD 13.4; N=237), while Sousa et al [59] mention an average SDMT performance score of 53.51 (SD 11.76; n=115). We found that the people with MS included in this study performed significantly better (comparison with López-Góngora et al [58]: F1,336=8.01; P=.005; comparison with Sousa et al [59]: F1,214=11.1; P=.001).

The underlying reason for having selected a cognitively preserved sample is most likely the sampling bias resulting from the limited inclusion and exclusion criteria and the recruitment of patients during outpatient consultation in secondary and tertiary MS care settings. In these settings, participants are tested cognitively at least annually, which might enhance their familiarity with typical cognitive tests for MS, with a carryover effect to derivative tests, such as those in icognition. Moreover, people with MS who were able and willing to participate might (1) have better cognitive abilities, (2) be better at handling a smartphone, and (3) be familiar with scientific studies. The reason for being less strict on inclusion was to be able to offer the screening battery to any patient with MS who is regularly followed up in an outpatient setting. Future studies should confirm the validity of icognition in people with MS who are cognitively impaired. Moreover, combined with the publicly available data underlying this study, the sensitivity of icognition in detecting cognitive impairment could be established.

To investigate the impact of cognitive impairment on our results, we performed a post hoc analysis excluding participants who were considered impaired on at least 1 paper-and-pencil test (Figures S3-S5 and Table S3 in Multimedia Appendix 1). Compared to the original analysis, the most notable differences were observed in the correlations with clinical parameters, where the correlation between the Symbol Test and disease duration became significant (P=.04), while significance dropped for the relationship between the vBDS and education level (P=.06), the Dot Test and FSMC (P=.18), and the vBDS and EDSS (P=.20). Results for concurrent validity, test-retest reliability, and the difference between people with MS and healthy controls were comparable.

The Benefits of Regular Digital Follow-Up

Proper and regular follow-up of cognitive function is important to capture fluctuations in cognitive state, such as those caused by a disease exacerbation or a relapse. Although Giedraitiene et al [60] show that a short cognitive screening tool such as the Brief International Cognitive Assessment for Multiple Sclerosis [18] can detect these cognitive fluctuations [60], the cognitive dip is likely missed in reality because cognitive assessments are usually performed only once or twice a year during clinical follow-up visits. The icognition screening battery allows for more frequent cognitive assessments. Furthermore, testing can be performed wherever and whenever a patient feels ready for it, reducing biasing effects such as fatigue [61]. Moreover, digitalization facilitates data collection and allows the extraction of more information from a cognitive test. As icognition associates every response with a 13-digit time stamp (with millisecond precision), cognitive fatigue or other performance metrics such as the “maximum gap time between correct responses” [39] can be tracked with high temporal resolution. For the Symbol Test, analogous examples are provided in the studies by Ganzetti et al [39] and Barrios et al [40].

Implementing icognition in a health care platform that includes fine motor assessment [62], symptom logging, and patient-reported outcome measures [21] enables patients to take an active role in their disease management; they can provide the individual information that complements the professional knowledge of the treating physician [63]. Moreover, collecting confounding variables, such as fatigue [21], at cognitive assessment improves the interpretation of cognitive scores.

Limitations and Future Work

This study has some limitations. This validation study used a cross-sectional design, aside from the assessment of test-retest reliability. Therefore, we were unable to map the cognitive evolution of people with MS over time. These and other validated digital cognitive tests could considerably facilitate longitudinal testing in future studies, analogous to the 1-year follow-up study by Lam et al [64].

Popular screening batteries such as the Brief International Cognitive Assessment for Multiple Sclerosis also include a measure of verbal memory, which is indeed commonly impaired in MS [14]. However, we chose not to implement a digital test of verbal memory in icognition because such tests usually rely heavily on language, which complicates international use with the need for language-specific validation. Moreover, this introduces new technological difficulties, such as variations in microphone and speaker quality or the accuracy of speech recognition software. Although the aim of this study was to create a quick, home-based cognitive screening assessment, we acknowledge the importance of more in-depth multidomain cognitive assessment, including the assessment of verbal memory.

In the initial version of the Dot Test, the final 5 trials included remembering the position of the 3 dots on a 5×5 grid, but, after testing 3 participants with MS, we decided to consistently use a 4×4 grid, given the difficulty of the task, and excluded these participants from the Dot Test analyses. However, including these 3 participants in a post hoc analysis yielded similar results.

Furthermore, we tested consistently on the same Android smartphone (Samsung Galaxy A10) to avoid a biasing influence of model-specific factors, such as screen size, weight, and responsiveness, because these factors might impact performance and user experience. Although we expect our results to be applicable to other smartphone models, this limits the study findings to this smartphone model. However, as icognition has been designed using the Flutter framework (Google LLC), it can also be deployed on operating systems other than Android. An external validation study is planned to show the robustness of the app on different devices in a home setting.

To test concurrent validity, we used the SPART as the ground truth test for visuospatial memory because this closely resembled the Dot Memory Test presented in the study by Sliwinski et al [24]. Although recent evidence suggests that the SPART is among the most sensitive memory assessments in MS [65], Strober et al [66] found the Brief Visuospatial Memory Test–Revised to be more sensitive than the SPART. One might argue that the sole criterion of the ground truth test is that it measures visuospatial memory, regardless of its similarity to the digital test. In this context, using the Brief Visuospatial Memory Test–Revised would have been justified. Although the concurrent validity of the Dot Test was only fair, it demonstrated an acceptable test-retest reliability and correlated with several clinical variables, underscoring its value for clinical practice. Unlike the Symbol Test, which can be seen as a digital alternative to its paper-and-pencil reference test (SDMT), more research is necessary to establish the ground truth reference for the Dot Test. Furthermore, future research could subject icognition to a real-world (referred to later in this subsection), longitudinal study to establish its sensitivity to change at the individual level.

Test-retest reliability was assessed only in the healthy controls as per the recommendations of Benedict et al [28], who stated that test-retest reliability “can be investigated in a small sample of MS and/or healthy volunteers over 1-3 weeks.” Although we deem this to be a good indication for test-retest reliability in people with MS, substantiated by similar ICC reliability scores in the smartphone SDMT study by van Oirschot et al [35], test-retest reliability is not guaranteed to be similar for the cognitive tests in icognition. As we strove to perform test-retest reliability in similar circumstances (eg, location), assessing it for people with MS was difficult because they were tested during a 1-time visit to the outpatient MS clinics for consultation. Furthermore, we acknowledge that, although we followed the recommendations of Benedict et al [28], the sample sizes for testing test-retest reliability were small (Symbol Test and vBDS: n=20; Dot Test: n=19).

By recruiting a cognitively preserved sample of people with MS, this study was unfortunately unable to provide a realistic assessment of the sensitivity of the app to discriminate between (1) people with MS and healthy controls and (2) people with MS who are cognitively impaired and people with MS who are cognitively preserved. As the goal of icognition is not the diagnosis of MS but to screen for people with MS who are cognitively impaired, we especially recognize the latter as a limitation of this study. Indeed, of the 101 participants with MS, only 7 (6.9%), 8 (7.9%), and 9 (8.9%) presented impairment on the Symbol Test, Dot Test, and vBDS, respectively, resulting in a substantial class imbalance with respect to preserved test performance. However, because we make our data publicly available, this analysis can be conducted by future studies evaluating icognition on a sample of people with MS who are cognitively impaired.

This study validated the icognition app in controlled laboratory conditions under the supervision of a test examiner. However, we acknowledge that future studies should confirm psychometric properties such as test-retest reliability in remote, unsupervised settings. Furthermore, as recently demonstrated by Podda et al [15], the usability of cognitive apps should be assessed to assure a smooth transition to a real-life setting. A usability study is planned for icognition as well.

Conclusions

This study established the reliability and validity of a newly developed smartphone app, icognition, for remotely screening cognitive impairment in terms of IPS and memory in people with MS. This allows for regular screening of cognitive performance to more quickly detect and respond to potential deterioration.

Acknowledgments

The authors would like to thank Steve De Backer for his efforts in the development of the icognition app, as well as Eva Keytsman and Florine Wöhler for their support in recruiting and testing participants for this study. SD was funded by a Baekeland grant from Flanders Innovation and Entrepreneurship (HBC.2019.2579), a grant from Vrije Universiteit Brussel (SRP85), and a junior postdoctoral grant from Fonds Wetenschappelijk Onderzoek–Flanders (12A6U25N). Both DVL (1SD5322N) and GN (1805620N) are supported by a personal grant from Fonds Wetenschappelijk Onderzoek–Flanders. This study is partly funded by the CLAIMS (Clinical Impact Through AI-Assisted MS Care) project, supported by the Innovative Health Initiative joint undertaking (101112153). The joint undertaking is supported by the European Union’s Horizon Europe research and innovation program as well as COCIR (European Coordination Committee of the Radiological, Electromedical, and Healthcare Information Technology Industry), the European Federation of Pharmaceutical Industries and Associations, EuropaBio, MedTech Europe, Vaccines Europe, AB Science SA, and icometrix NV.

Abbreviations

BDI

Beck Depression Inventory

EDSS

Expanded Disability Status Scale

FSMC

Fatigue Scale for Motor and Cognitive Functions

ICC

intraclass correlation coefficient

IPS

information processing speed

MS

multiple sclerosis

SDMT

Symbol Digit Modalities Test

SPART

10/36 Spatial Recall Test

vBDS

visual backward digit span

Multimedia Appendix 1

Supplementary text, tables, and figures.

jmir_v27i1e53503_app1.docx (790.5KB, docx)

Data Availability

The anonymized source data on which this paper relies, as well as the code used for statistical analysis and the creation of figures, are available in the GitHub repository of our laboratory [67].

Footnotes

Conflicts of Interest: This work was part of the industrial PhD trajectory of SD in collaboration with icometrix. LC, AD, DS, and DMS are employees of icometrix. GN and DS are shareholders of icometrix. All other authors declare no other conflicts of interest.

References

  • 1.Lee S, Kim HS. Prospect of artificial intelligence based on electronic medical record. J Lipid Atheroscler. 2021 Sep;10(3):282–90. doi: 10.12997/jla.2021.10.3.282. https://europepmc.org/abstract/MED/34621699 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Mohsen F, Ali H, El Hajj N, Shah Z. Artificial intelligence-based methods for fusion of electronic health records and imaging data. Sci Rep. 2022 Oct 26;12(1):17981. doi: 10.1038/s41598-022-22514-4. https://doi.org/10.1038/s41598-022-22514-4 .10.1038/s41598-022-22514-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Getachew E, Adebeta T, Muzazu SG, Charlie L, Said B, Tesfahunei HA, Wanjiru CL, Acam J, Kajogoo VD, Solomon S, Atim MG, Manyazewal T. Digital health in the era of COVID-19: reshaping the next generation of healthcare. Front Public Health. 2023;11:942703. doi: 10.3389/fpubh.2023.942703. https://europepmc.org/abstract/MED/36875401 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Jakimovski D, Bittner S, Zivadinov R, Morrow SA, Benedict RH, Zipp F, Weinstock-Guttman B. Multiple sclerosis. Lancet. 2024 Jan 13;403(10422):183–202. doi: 10.1016/S0140-6736(23)01473-3.S0140-6736(23)01473-3 [DOI] [PubMed] [Google Scholar]
  • 5.Yeroushalmi S, Maloni H, Costello K, Wallin MT. Telemedicine and multiple sclerosis: a comprehensive literature review. J Telemed Telecare. 2020;26(7-8):400–13. doi: 10.1177/1357633X19840097. [DOI] [PubMed] [Google Scholar]
  • 6.Khan F, Amatya B, Kesselring J, Galea M. Telerehabilitation for persons with multiple sclerosis. Cochrane Database Syst Rev. 2015 Apr 09;2015(4):CD010508. doi: 10.1002/14651858.CD010508.pub2. https://europepmc.org/abstract/MED/25854331 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Charvet LE, Yang J, Shaw MT, Sherman K, Haider L, Xu J, Krupp LB. Cognitive function in multiple sclerosis improves with telerehabilitation: results from a randomized controlled trial. PLoS One. 2017;12(5):e0177177. doi: 10.1371/journal.pone.0177177. https://dx.plos.org/10.1371/journal.pone.0177177 .PONE-D-16-46905 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Ruano L, Portaccio E, Goretti B, Niccolai C, Severo M, Patti F, Cilia S, Gallo P, Grossi P, Ghezzi A, Roscio M, Mattioli F, Stampatori C, Trojano M, Viterbo RG, Amato MP. Age and disability drive cognitive impairment in multiple sclerosis across disease subtypes. Mult Scler. 2017 Aug 13;23(9):1258–67. doi: 10.1177/1352458516674367.1352458516674367 [DOI] [PubMed] [Google Scholar]
  • 9.DeLuca J, Chiaravalloti ND, Sandroff BM. Treatment and management of cognitive dysfunction in patients with multiple sclerosis. Nat Rev Neurol. 2020 Jun 05;16(6):319–32. doi: 10.1038/s41582-020-0355-1.10.1038/s41582-020-0355-1 [DOI] [PubMed] [Google Scholar]
  • 10.Pozzilli V, Cruciani A, Capone F, Motolese F, Rossi M, Pilato F, Bianco A, Mirabella M, Giuffré GM, Calabria LF, Di Lazzaro V. A true isolated cognitive relapse in multiple sclerosis. Neurol Sci. 2023 Jan;44(1):339–42. doi: 10.1007/s10072-022-06441-w.10.1007/s10072-022-06441-w [DOI] [PubMed] [Google Scholar]
  • 11.Benedict RH, Morrow S, Rodgers J, Hojnacki D, Bucello MA, Zivadinov R, Weinstock-Guttman B. Characterizing cognitive function during relapse in multiple sclerosis. Mult Scler. 2014 Nov;20(13):1745–52. doi: 10.1177/1352458514533229.1352458514533229 [DOI] [PubMed] [Google Scholar]
  • 12.Global smartphone penetration rate as share of population from 2016 to 2023. Statista. [2024-05-23]. https://www.statista.com/statistics/203734/global-smartphone-penetration-per-capita-since-2005/
  • 13.Foong YC, Bridge F, Merlo D, Gresle M, Zhu C, Buzzard K, Butzkueven H, van der Walt A. Smartphone monitoring of cognition in people with multiple sclerosis: a systematic review. Mult Scler Relat Disord. 2023 May;73:104674. doi: 10.1016/j.msard.2023.104674. https://linkinghub.elsevier.com/retrieve/pii/S2211-0348(23)00178-5 .S2211-0348(23)00178-5 [DOI] [PubMed] [Google Scholar]
  • 14.Macías Islas MÁ, Ciampi E. Assessment and impact of cognitive impairment in multiple sclerosis: an overview. Biomedicines. 2019 Mar 19;7(1):22. doi: 10.3390/biomedicines7010022. https://www.mdpi.com/resolver?pii=biomedicines7010022 .biomedicines7010022 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Podda J, Tacchino A, Ponzio M, Di Antonio F, Susini A, Pedullà L, Battaglia MA, Brichetto G. Mobile health app (DIGICOG-MS) for self-assessment of cognitive impairment in people with multiple sclerosis: instrument validation and usability study. JMIR Form Res. 2024 Jun 20;8:e56074. doi: 10.2196/56074. https://formative.jmir.org/2024//e56074/ v8i1e56074 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Beier M, Alschuler K, Amtmann D, Hughes A, Madathil R, Ehde D. iCAMS: assessing the reliability of a brief international cognitive assessment for multiple sclerosis (BICAMS) tablet application. Int J MS Care. 2020;22(2):67–74. doi: 10.7224/1537-2073.2018-108. https://europepmc.org/abstract/MED/32410901 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Benedict RH, DeLuca J, Phillips G, LaRocca N, Hudson LD, Rudick R, Consortium MS. Validity of the symbol digit modalities test as a cognition performance outcome measure for multiple sclerosis. Mult Scler. 2017 Apr;23(5):721–33. doi: 10.1177/1352458517690821. https://journals.sagepub.com/doi/abs/10.1177/1352458517690821?url_ver=Z39.88-2003&rfr_id=ori:rid:crossref.org&rfr_dat=cr_pub0pubmed . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Langdon DW, Amato MP, Boringa J, Brochet B, Foley F, Fredrikson S, Hämäläinen P, Hartung HP, Krupp L, Penner IK, Reder AT, Benedict RH. Recommendations for a brief international cognitive assessment for multiple sclerosis (BICAMS) Mult Scler. 2012 Jun;18(6):891–8. doi: 10.1177/1352458511431076. https://journals.sagepub.com/doi/abs/10.1177/1352458511431076?url_ver=Z39.88-2003&rfr_id=ori:rid:crossref.org&rfr_dat=cr_pub0pubmed .1352458511431076 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Pereira DR, Costa P, Cerqueira JJ. Repeated assessment and practice effects of the written symbol digit modalities test using a short inter-test interval. Arch Clin Neuropsychol. 2015 Aug;30(5):424–34. doi: 10.1093/arclin/acv028.acv028 [DOI] [PubMed] [Google Scholar]
  • 20.Lamers I, Kerkhofs L, Raats J, Kos D, van Wijmeersch B, Feys P. Perceived and actual arm performance in multiple sclerosis: relationship with clinical tests according to hand dominance. Mult Scler. 2013 Sep;19(10):1341–8. doi: 10.1177/1352458513475832.1352458513475832 [DOI] [PubMed] [Google Scholar]
  • 21.van Hecke W, Costers L, Descamps A, Ribbens A, Nagels G, Smeets D, Sima DM. A novel digital care management platform to monitor clinical and subclinical disease activity in multiple sclerosis. Brain Sci. 2021 Sep 03;11(9):1171. doi: 10.3390/brainsci11091171. https://www.mdpi.com/resolver?pii=brainsci11091171 .brainsci11091171 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Thompson AJ, Banwell BL, Barkhof F, Carroll WM, Coetzee T, Comi G, Correale J, Fazekas F, Filippi M, Freedman MS, Fujihara K, Galetta SL, Hartung HP, Kappos L, Lublin FD, Marrie RA, Miller AE, Miller DH, Montalban X, Mowry EM, Sorensen PS, Tintoré M, Traboulsee AL, Trojano M, Uitdehaag BM, Vukusic S, Waubant E, Weinshenker BG, Reingold SC, Cohen JA. Diagnosis of multiple sclerosis: 2017 revisions of the McDonald criteria. Lancet Neurol. 2018 Feb;17(2):162–73. doi: 10.1016/S1474-4422(17)30470-2.S1474-4422(17)30470-2 [DOI] [PubMed] [Google Scholar]
  • 23.Rypma B, Berger JS, Prabhakaran V, Bly BM, Kimberg DY, Biswal BB, D'Esposito M. Neural correlates of cognitive efficiency. Neuroimage. 2006 Nov 15;33(3):969–79. doi: 10.1016/j.neuroimage.2006.05.065.S1053-8119(06)00605-7 [DOI] [PubMed] [Google Scholar]
  • 24.Sliwinski MJ, Mogle JA, Hyun J, Munoz E, Smyth JM, Lipton RB. Reliability and validity of ambulatory cognitive assessments. Assessment. 2018 Jan;25(1):14–30. doi: 10.1177/1073191116643164. https://europepmc.org/abstract/MED/27084835 .1073191116643164 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Hilbert S, Nakagawa TT, Puci P, Zech A, Bühner M. The digit span backwards task: verbal and visual cognitive strategies in working memory assessment. Eur J Psychol Assess. 2015;31(3):174–80. doi: 10.1027/1015-5759/A000223. [DOI] [Google Scholar]
  • 26.Woods DL, Herron TJ, Yund EW, Hink RF, Kishiyama MM, Reed B. Computerized analysis of error patterns in digit span recall. J Clin Exp Neuropsychol. 2011 Aug;33(7):721–34. doi: 10.1080/13803395.2010.550602. [DOI] [PubMed] [Google Scholar]
  • 27.Pham L, Harris T, Varosanec M, Morgan V, Kosa P, Bielekova B. Smartphone-based symbol-digit modalities test reliably captures brain damage in multiple sclerosis. NPJ Digit Med. 2021 Feb 24;4(1):36. doi: 10.1038/s41746-021-00401-y. https://doi.org/10.1038/s41746-021-00401-y .10.1038/s41746-021-00401-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Benedict RH, Amato MP, Boringa J, Brochet B, Foley F, Fredrikson S, Hamalainen P, Hartung H, Krupp L, Penner I, Reder AT, Langdon D. Brief International Cognitive Assessment for MS (BICAMS): international standards for validation. BMC Neurol. 2012 Jul 16;12(1):55. doi: 10.1186/1471-2377-12-55. https://bmcneurol.biomedcentral.com/articles/10.1186/1471-2377-12-55 .1471-2377-12-55 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Gerstenecker A, Martin R, Marson DC, Bashir K, Triebel KL. Introducing demographic corrections for the 10/36 spatial recall test. Int J Geriatr Psychiatry. 2016 Apr;31(4):406–11. doi: 10.1002/gps.4346. https://europepmc.org/abstract/MED/26270773 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Raiford SE, Coalson DL, Saklofske DH, Weiss LG. Practical issues in WAIS-IV administration and scoring. In: Weiss LG, Saklofske DH, Coalson DL, Raiford SE, editors. WAIS-IV Clinical Use and Interpretation: A Volume in Practical Resources for the Mental Health Professional. New York, NY: Academic Press; 2010. pp. 25–59. [Google Scholar]
  • 31.Woods DL, Kishiyamaa MM, Lund EW, Herron TJ, Edwards B, Poliva O, Hink RF, Reed B. Improving digit span assessment of short-term verbal memory. J Clin Exp Neuropsychol. 2011 Jan;33(1):101–11. doi: 10.1080/13803395.2010.493149. https://europepmc.org/abstract/MED/20680884 .925010448 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Kurtzke JF. Rating neurologic impairment in multiple sclerosis: an expanded disability status scale (EDSS) Neurology. 1983 Nov;33(11):1444–52. doi: 10.1212/wnl.33.11.1444. [DOI] [PubMed] [Google Scholar]
  • 33.Beck AT, Ward CH, Mendelson M, Mock J, Erbaugh J. An inventory for measuring depression. Arch Gen Psychiatry. 1961 Jun;4:561–71. doi: 10.1001/archpsyc.1961.01710120031004. [DOI] [PubMed] [Google Scholar]
  • 34.Penner IK, Raselli C, Stöcklin M, Opwis K, Kappos L, Calabrese P. The Fatigue Scale for Motor and Cognitive functions (FSMC): validation of a new instrument to assess multiple sclerosis-related fatigue. Mult Scler. 2009 Dec;15(12):1509–17. doi: 10.1177/1352458509348519.1352458509348519 [DOI] [PubMed] [Google Scholar]
  • 35.van Oirschot P, Heerings M, Wendrich K, den Teuling B, Martens MB, Jongen PJ. Symbol digit modalities test variant in a smartphone app for persons with multiple sclerosis: validation study. JMIR Mhealth Uhealth. 2020 Oct 05;8(10):e18160. doi: 10.2196/18160. https://mhealth.jmir.org/2020/10/e18160/ v8i10e18160 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Koo TK, Li MY. A guideline of selecting and reporting intraclass correlation coefficients for reliability research. J Chiropr Med. 2016 Jun;15(2):155–63. doi: 10.1016/j.jcm.2016.02.012. https://europepmc.org/abstract/MED/27330520 .S1556-3707(16)00015-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Portney LG, Watkins MP. Foundations of Clinical Research: Applications to Practice. London, UK: Pearson/Prentice Hall; 2009. [Google Scholar]
  • 38.Benedict RH, Cookfair D, Gavett R, Gunther M, Munschauer F, Garg N, Weinstock-Guttman B. Validity of the minimal assessment of cognitive function in multiple sclerosis (MACFIMS) J Int Neuropsychol Soc. 2006 Jul 27;12(4):549–58. doi: 10.1017/s1355617706060723. [DOI] [PubMed] [Google Scholar]
  • 39.Ganzetti M, Graves JS, Holm SP, Dondelinger F, Midaglia L, Gaetano L, Craveiro L, Lipsmeier F, Bernasconi C, Montalban X, Hauser SL, Lindemann M. Neural correlates of digital measures shown by structural MRI: a post-hoc analysis of a smartphone-based remote assessment feasibility study in multiple sclerosis. J Neurol. 2023 Mar 05;270(3):1624–36. doi: 10.1007/s00415-022-11494-0. https://europepmc.org/abstract/MED/36469103 .10.1007/s00415-022-11494-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Barrios L, Amon R, Oldrati P, Hilty M, Holz C, Lutterotti A. Cognitive fatigability assessment test (cFAST): development of a new instrument to assess cognitive fatigability and pilot study on its association to perceived fatigue in multiple sclerosis. Digit Health. 2022;8:20552076221117740. doi: 10.1177/20552076221117740. https://journals.sagepub.com/doi/10.1177/20552076221117740?url_ver=Z39.88-2003&rfr_id=ori:rid:crossref.org&rfr_dat=cr_pub0pubmed .10.1177_20552076221117740 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Chen MH, Cherian C, Elenjickal K, Rafizadeh CM, Ross MK, Leow A, DeLuca J. Real-time associations among MS symptoms and cognitive dysfunction using ecological momentary assessment. Front Med (Lausanne) 2022;9:1049686. doi: 10.3389/fmed.2022.1049686. https://europepmc.org/abstract/MED/36714150 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Weinstock ZL, Jaworski M, Dwyer MG, Jakimovski D, Burnham A, Wicks TR, Youngs M, Santivasci C, Cruz S, Gillies J, Covey TJ, Suchan C, Bergsland N, Weinstock-Guttman B, Zivadinov R, Benedict RH. Auditory test of processing speed: preliminary validation of a smartphone-based test of mental speed. Mult Scler. 2023 Nov;29(13):1646–58. doi: 10.1177/13524585231199311. [DOI] [PubMed] [Google Scholar]
  • 43.Lam KH, Twose J, Lissenberg-Witte B, Licitra G, Meijer K, Uitdehaag B, de Groot V, Killestein J. The use of smartphone keystroke dynamics to passively monitor upper limb and cognitive function in multiple sclerosis: longitudinal analysis. J Med Internet Res. 2022 Nov 07;24(11):e37614. doi: 10.2196/37614. https://www.jmir.org/2022/11/e37614/ v24i11e37614 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Chen MH, Leow A, Ross MK, DeLuca J, Chiaravalloti N, Costa SL, Genova HM, Weber E, Hussain F, Demos AP. Associations between smartphone keystroke dynamics and cognition in MS. Digit Health. 2022 Dec 05;8:20552076221143234. doi: 10.1177/20552076221143234. https://journals.sagepub.com/doi/10.1177/20552076221143234?url_ver=Z39.88-2003&rfr_id=ori:rid:crossref.org&rfr_dat=cr_pub0pubmed .10.1177_20552076221143234 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Pless S, Woelfle T, Naegelin Y, Lorscheider J, Wiencierz A, Reyes Ó, Calabrese P, Kappos L. Assessment of cognitive performance in multiple sclerosis using smartphone-based training games: a feasibility study. J Neurol. 2023 Jul;270(7):3451–63. doi: 10.1007/s00415-023-11671-9. https://europepmc.org/abstract/MED/36952010 .10.1007/s00415-023-11671-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Maillart E, Labauge P, Cohen M, Maarouf A, Vukusic S, Donzé C, Gallien P, de Sèze J, Bourre B, Moreau T, Louapre C, Mayran P, Bieuvelet S, Vallée M, Bertillot F, Klaeylé L, Argoud A, Zinaï S, Tourbah A. MSCopilot, a new multiple sclerosis self-assessment digital solution: results of a comparative study versus standard tests. Eur J Neurol. 2020 Mar;27(3):429–36. doi: 10.1111/ene.14091. [DOI] [PubMed] [Google Scholar]
  • 47.Pratap A, Grant D, Vegesna A, Tummalacherla M, Cohan S, Deshpande C, Mangravite L, Omberg L. Evaluating the utility of smartphone-based sensor assessments in persons with multiple sclerosis in the real-world using an app (elevateMS): observational, prospective pilot digital health study. JMIR Mhealth Uhealth. 2020 Oct 27;8(10):e22108. doi: 10.2196/22108. https://mhealth.jmir.org/2020/10/e22108/ v8i10e22108 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Ross MK, Demos AP, Zulueta J, Piscitello A, Langenecker SA, McInnis M, Ajilore O, Nelson PC, Ryan KA, Leow A. Naturalistic smartphone keyboard typing reflects processing speed and executive function. Brain Behav. 2021 Nov;11(11):e2363. doi: 10.1002/brb3.2363. https://europepmc.org/abstract/MED/34612605 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Montalban X, Graves J, Midaglia L, Mulero P, Julian L, Baker M, Schadrack J, Gossens C, Ganzetti M, Scotland A, Lipsmeier F, van Beek J, Bernasconi C, Belachew S, Lindemann M, Hauser SL. A smartphone sensor-based digital outcome assessment of multiple sclerosis. Mult Scler. 2022 Apr;28(4):654–64. doi: 10.1177/13524585211028561. https://journals.sagepub.com/doi/abs/10.1177/13524585211028561?url_ver=Z39.88-2003&rfr_id=ori:rid:crossref.org&rfr_dat=cr_pub0pubmed . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Woelfle T, Pless S, Wiencierz A, Kappos L, Naegelin Y, Lorscheider J. Practice effects of mobile tests of cognition, dexterity, and mobility on patients with multiple sclerosis: data analysis of a smartphone-based observational study. J Med Internet Res. 2021 Nov 18;23(11):e30394. doi: 10.2196/30394. https://www.jmir.org/2021/11/e30394/ v23i11e30394 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Metzger R, Garrett K, Christensen A, Foley J. Digital performance measures show sensitivity to demographic and disease characteristics in a multiple sclerosis cohort utilizing the MS care connect mobile app in a real-world setting (P18-4.004) Neurology. 2022 May 03;98(18_supplement):11. doi: 10.1212/wnl.98.18_supplement.2053. [DOI] [Google Scholar]
  • 52.Michaud JB, Penny C, Cull O, Hervet E, Chamard-Witkowski L. Remote testing apps for multiple sclerosis patients: scoping review of published articles and systematic search and review of public smartphone apps. JMIR Neurotech. 2023 Feb 6;2:e37944. doi: 10.2196/37944. [DOI] [Google Scholar]
  • 53.MS Care Connect. [2024-05-29]. https://www.mscareconnect.com/
  • 54.Scaramozza M, Chiesa PA, Zajac L, Sun Z, Tang M, Juraver A, Bartholomé E, Charré-Morin J, Saubusse A, Johnson SC, Brochet B, Carment L, Ruiz M, Campbell N, Ruet A. Konectom™ cognitive processing speed test enables reliable remote, unsupervised cognitive assessment in people with multiple sclerosis: exploring the use of substitution time as a novel digital outcome measure. Mult Scler. 2024 Aug;30(9):1193–204. doi: 10.1177/13524585241259650. [DOI] [PubMed] [Google Scholar]
  • 55.van der Feen FE, de Haan GA, van der Lijn I, Huizinga F, Meilof JF, Heersema DJ, Heutink J. Recognizing visual complaints in people with multiple sclerosis: prevalence, nature and associations with key characteristics of MS. Mult Scler Relat Disord. 2022 Jan;57:103429. doi: 10.1016/j.msard.2021.103429. https://linkinghub.elsevier.com/retrieve/pii/S2211-0348(21)00695-7 .S2211-0348(21)00695-7 [DOI] [PubMed] [Google Scholar]
  • 56.Ghandi Dezfuli M, Akbarfahimi M, Nabavi SM, Hassani Mehraban A, Jafarzadehpur E. Can hand dexterity predict the disability status of patients with multiple sclerosis? Med J Islam Repub Iran. 2015;29:255. https://europepmc.org/abstract/MED/26793646 . [PMC free article] [PubMed] [Google Scholar]
  • 57.Anders K. Resolution of students t-tests, ANOVA and analysis of variance components from intermediary data. Biochem Med (Zagreb) 2017 Jun 15;27(2):253–8. doi: 10.11613/BM.2017.026. https://europepmc.org/abstract/MED/28740445 .bm-27-253 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.López-Góngora M, Querol L, Escartín A. A one-year follow-up study of the Symbol Digit Modalities Test (SDMT) and the Paced Auditory Serial Addition Test (PASAT) in relapsing-remitting multiple sclerosis: an appraisal of comparative longitudinal sensitivity. BMC Neurol. 2015 Mar 22;15(1):40. doi: 10.1186/s12883-015-0296-2. https://bmcneurol.biomedcentral.com/articles/10.1186/s12883-015-0296-2 .10.1186/s12883-015-0296-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Sousa C, Rigueiro-Neves M, Passos AM, Ferreira A, Sá MJ, Group for Validation of the BRBN-T in the Portuguese MS Population Assessment of cognitive functions in patients with multiple sclerosis applying the normative values of the Rao's brief repeatable battery in the Portuguese population. BMC Neurol. 2021 Apr 21;21(1):170. doi: 10.1186/s12883-021-02193-w. https://bmcneurol.biomedcentral.com/articles/10.1186/s12883-021-02193-w .10.1186/s12883-021-02193-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Giedraitiene N, Kaubrys G, Kizlaitiene R. Cognition during and after multiple sclerosis relapse as assessed with the brief international cognitive assessment for multiple sclerosis. Sci Rep. 2018 May 25;8(1):8169. doi: 10.1038/s41598-018-26449-7. https://doi.org/10.1038/s41598-018-26449-7 .10.1038/s41598-018-26449-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Andreasen AK, Iversen P, Marstrand L, Siersma V, Siebner HR, Sellebjerg F. Structural and cognitive correlates of fatigue in progressive multiple sclerosis. Neurol Res. 2019 Feb;41(2):168–76. doi: 10.1080/01616412.2018.1547813. [DOI] [PubMed] [Google Scholar]
  • 62.van Laethem D, Denissen S, Costers L, Descamps A, Baijot J, van Remoortel A, van Merhaegen-Wieleman A, D'hooghe MB, D'Haeseleer M, Smeets D, Sima D, van Schependom J, Nagels G. The finger dexterity test: validation study of a smartphone-based manual dexterity assessment. Mult Scler. 2024 Jan;30(1):121–30. doi: 10.1177/13524585231216007. [DOI] [PubMed] [Google Scholar]
  • 63.Holman H, Lorig K. Patients as partners in managing chronic disease. Partnership is a prerequisite for effective and efficient health care. BMJ. 2000 Feb 26;320(7234):526–7. doi: 10.1136/bmj.320.7234.526. https://europepmc.org/abstract/MED/10688539 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Lam KH, Bucur IG, van Oirschot PV, de Graaf FD, Weda H, Strijbis E, Uitdehaag B, Heskes T, Killestein J, de Groot V. Towards individualized monitoring of cognition in multiple sclerosis in the digital era: a one-year cohort study. Mult Scler Relat Disord. 2022 Apr;60:103692. doi: 10.1016/j.msard.2022.103692. https://linkinghub.elsevier.com/retrieve/pii/S2211-0348(22)00207-3 .S2211-0348(22)00207-3 [DOI] [PubMed] [Google Scholar]
  • 65.Raimo S, Giorgini R, Gaita M, Costanzo A, Spitaleri D, Palermo L, Liuzza MT, Santangelo G. Sensitivity of conventional cognitive tests in multiple sclerosis: application of item response theory. Mult Scler Relat Disord. 2023 Jan;69:104440. doi: 10.1016/j.msard.2022.104440.S2211-0348(22)00944-0 [DOI] [PubMed] [Google Scholar]
  • 66.Strober L, Englert J, Munschauer F, Weinstock-Guttman B, Rao S, Benedict R. Sensitivity of conventional memory tests in multiple sclerosis: comparing the Rao Brief Repeatable Neuropsychological Battery and the Minimal Assessment of Cognitive Function in MS. Mult Scler. 2009 Sep;15(9):1077–84. doi: 10.1177/1352458509106615.1352458509106615 [DOI] [PubMed] [Google Scholar]
  • 67.AIMS-VUB / smartphone_tests. GitHub. [2024-12-30]. https://github.com/AIMS-VUB/smartphone_tests .

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Multimedia Appendix 1

Supplementary text, tables, and figures.

jmir_v27i1e53503_app1.docx (790.5KB, docx)

Data Availability Statement

The anonymized source data on which this paper relies, as well as the code used for statistical analysis and the creation of figures, are available in the GitHub repository of our laboratory [67].


Articles from Journal of Medical Internet Research are provided here courtesy of JMIR Publications Inc.

RESOURCES