Skip to main content
Springer logoLink to Springer
. 2023 Mar 13;34(2):363–401. doi: 10.1007/s11065-023-09586-3

The Neuropsychological Assessment of Unilateral Spatial Neglect Through Computerized and Virtual Reality Tools: A Scoping Review

Stefano Terruzzi 1,2,3,, Federica Albini 1,2, Gemma Massetti 2,4, Roberta Etzi 2, Alberto Gallace 1,2, Giuseppe Vallar 1,2,5,
PMCID: PMC10009867  PMID: 36913099

Abstract

Unilateral Spatial Neglect is a disabling neuropsychological deficit. Patients with spatial neglect fail to detect and report events, and to perform actions in the side of space contralateral to a hemispheric cerebral lesion. Neglect is assessed by evaluating the patients’ abilities in daily life activities and by psychometric tests. Computer-based, portable and Virtual Reality technologies may provide more and precise data, and be more sensitive and informative, compared to current paper-and-pencil procedures. Studies since 2010, in which such technologies have been used, are reviewed. Forty-two articles meeting inclusion criteria are categorized according to their technological approaches (computer-, graphics tablet or tablet-, virtual reality-based assessment, and other). The results are promising. However, a definite golden standard, technologically based procedure cannot be still established. Developing technologically based tests is a laborious process, which requires technical and user experience improvements as well as normative data, to increase the evidence of efficacy for clinical evaluation of at least some of the tests considered in this review.

Keywords: Information technology, Digital, Neuropsychological assessment, Unilateral spatial neglect

Introduction

Unilateral spatial neglect (henceforth spatial neglect) is a disabling neurological disorder. Patients fail to report events (in different sensory modalities: visual, auditory, and somatosensory; when imaging a familiar visual scene from a given vantage point), which occur in the side of physical and imaginal space and the body contralateral to the side of the hemispheric lesion (contralesional). These patients also fail to perform actions in those sides of extra-personal and personal (i.e., the body) spaces (Chokron et al., 2008; Heilman et al., 2000; Vallar & Bolognini, 2014; Vallar & Calzolari, 2018; Vallar & Maravita, 2009). In most patients, neglect impairs the left side of the space, contralateral to a right-hemispheric lesion. Right spatial neglect, contralateral to a left-hemispheric lesion, is less frequent and severe (Bisiach et al., 1984; Bowen et al., 1999; De Renzi et al., 1970; Ogden, 1985). The reported incidence of spatial neglect after stroke varies across studies (ranging from about 10 to 90%). This wide range depends on a number of factors, including the operational definition of neglect, the patients’ selection criteria, and the assessment methods (Bowen et al., 2013). Spatial neglect is more frequent in the acute stage post-stroke (Bowen et al., 1999). After about three months, spatial neglect is still present in over 15% of right-brain-damaged patients (Ringman et al., 2004). In stroke patients the presence of spatial neglect predicts poor outcome in functional recovery (Di Monaco et al., 2011), entailing a longer time of hospitalization, greater functional dependency, increased risk of falls and a long-term disability in Activities of Daily Living (ADL) (Barer, 1990; Batchelor et al., 2012; Bernspång et al., 1987; Bowen et al., 2013; Czernuszenko & Czlonkowska, 2009; Nijboer et al., 2013). This results in an overall increase of costs for the health system (Barrett & Houston, 2019).

Spatial neglect is not due to lower-level sensorimotor deficits. Many patients show associated sensorimotor deficits, such as hemiplegia, hemianesthesia and hemianopia. However, a double dissociation (Teuber, 1955; Vallar, 2000) has been found between spatial neglect and sensorimotor deficits: patients may show sensorimotor deficits without neglect which, in turn, may be present in isolation (Bisiach et al., 1986). However, sensory and motor deficits may have higher-order components, brought about by spatial neglect itself (Vallar et al., 1991; Vallar, Sandroni, Vallar et al., 1991a, b). In line with these behavioral findings, the neural correlates of spatial neglect involve damage to higher-order associative areas, in perisylvian regions, including the inferior parietal lobule, the temporo-parietal junction, the posterior part of the superior temporal gyrus, the premotor cortex, white matter fiber tracts connecting these regions, the thalamus and the basal ganglia (Corbetta & Shulman, 2011; Karnath & Rorden, 2012; Ptak & Schnider, 2011; Vallar & Calzolari, 2018).

Spatial neglect may be conceived as a multifaceted syndrome with many clinical manifestations that frequently co-occur, but may manifest also in isolation (Halligan et al., 2003; Vallar, 1998). A primary distinction may be drawn between two sets of abnormalities: defective and productive. Defective signs, in which patients exhibit the absence of the appropriate behavior, requested by the experimental task or the ADLs, are the most investigated components of the syndrome. They include the impairment at attending to and exploring the left side of space, and of objects in it (Albert, 1973; Bisiach et al., 1981, 1983; Bisiach & Vallar, 2000; Gainotti et al., 1972; Rode et al., 1995; Schenkenberg et al., 1980; Stone et al., 1991; Vallar & Perani, 1986). Patients with spatial neglect may also show productive manifestations, that consist in gratuitous actions and delusional beliefs. These behaviors are inappropriate in the clinical setting, with respect to the requests by the examiner (as for extra-personal space: Kleinman et al., 2013; Ronchi et al., 2012, 2016; Rusconi et al., 2002; as for bodily space: Vallar & Bolognini, 2014; Vallar & Ronchi, 2009). Spatial neglect may occur in different dimensions or frames, with reference to the body and body parts (the mid-sagittal plane of the trunk, the head). In egocentric spatial neglect, the neglected side is coded with reference to the body’s trunk, the head, or the eyes. Allocentric spatial neglect refers instead to one side of an object, encoded with reference to its intrinsic coordinates (Howard, 1982; Klatzky, 1998, about these reference frames; see Vallar, 1998; Vallar & Bolognini, 2014 for these different manifestations of neglect). Furthermore, depending on the affected sector of space, spatial neglect may concern personal (the body), near peri-personal (the space surrounding the body, and within the reach of body effectors, primarily the hand, also the foot), and far extra-personal (not within the reach of such effectors) sectors of space (Vallar & Maravita, 2009). Finally, spatial neglect may also concern internally generated visual images (representational or imaginal spatial neglect, see Bisiach & Luzzatti, 1978; Bisiach et al., 1981; Meador et al., 1987; Grossi et al., 1989; Rode et al., 1995). All these manifestations of spatial neglect may occur in association, but dissociations among virtually all signs of the deficit have been observed (Bisiach et al., 1986; Guariglia et al., 1993).

Quantitative and Standardized Assessment (“paper-and-pencil”) with Available Normative data

Up to the mid 1930’s (Weisenberg & McBride, 1935), the neuropsychological method coupled a qualitative, often insightful, analysis of the patient’s deficits with, whenever possible, a post-mortem examination of the patient’s brain, to determine the localization of the cerebral lesion, and to relate it with the observed symptoms and signs. Starting from the seminal observations by Broca about aphasia (Broca, 1861; Vallar & Caputi, 2021), the assessment of the patients’ behavioral deficits was qualitative, and consisted in the accurate observation of the behavioral deficits, which were typically apparent and, due to some peculiar features, had captured the attention of the examiner. The neuropsychological examination became then more and more based on standardized tests, which provide quantitative measures (i.e., scores) of the patients’ performances. Moreover, the performances of brain-damaged patients started to be compared, with the support of statistical procedures, with those of control participants, comparable to patients for demographic and socio-cultural variables, and differing from them only for the presence of the cerebral lesion and the putative deficit under investigation. The research work by the Italian neurologist Ennio De Renzi provides an example of this approach (Arrigoni & De Renzi, 1964; De Renzi, 1982). As to spatial neglect, the quantitative assessment of its clinical manifestations has traditionally relied on paper-and-pencil tests, which includes the tasks described in the following.

1) Target cancellation: to find out and mark targets printed on a sheet of paper. In the Line (Albert, 1973) and Circle (Bisiach et al., 1979; Vallar & Perani, 1986) tests only targets are displayed. In the Bells (Gauthier et al., 1989), Star (Wilson et al., 1987), and Letter (Diller & Weinberg, 1977) tests, targets are shown intermingled with distractors. Tests requiring target vs. non-target discrimination are more likely to detect spatial neglect than those in which only targets are present (Arduino et al., 2002; Vallar, 2000). Cancellation tasks, in which patients are required to discriminate between complete and incomplete targets (i.e., the incomplete ones with a left-sided or right-sided missing portion), allow to distinguish between egocentric vs. allocentric, stimulus-based, types of spatial neglect. In target cancellation tests, when left spatial neglect occurs in an egocentric reference frame (e.g., the mid-sagittal plane of the body), complete and incomplete targets, left-sided with reference to such a frame, are not crossed out. In allocentric, object-based left spatial neglect, targets with a left-sided missing part are not crossed out across the whole display, with reference to coordinates based on the target itself (Bickerton et al., 2011; Mancuso et al., 2015; Ota et al., 2001, 2003). Quantitative scores include the total number of omissions and the difference between the number of omissions in the two sides (left and right) of the display (a sheet of paper). Moreover, these tests may provide qualitative information also about other aspects of the patients’ performance, including: (i) the location of the first target crossed out; (ii) the directional pattern of exploration; (iii) the speed of execution, and (iv) the types of error. Errors include the defective omission of targets, that are not crossed out, and the productive manifestations: simple perseveration (repeated crossing out marks), and complex gratuitous productions (various types of drawings, e.g., a hen, the patients’ signature) (Gandola et al., 2013; Ronchi et al., 2009; Rusconi et al., 2002). Another measure of spatial neglect in cancellation tasks is the “center of cancellation”, consisting in the scaled mean position in the horizontal (left-right) dimension of cancelled targets (Rorden & Karnath, 2010; Toraldo et al., 2017).

2) Bisection of a horizontal line: to mark the location corresponding to the physical mid-point of a line, which is then divided it into two segments. In the standard version of line bisection patients set the mid-point using the not paretic arm and hand, ipsilateral to the side of the lesion (right-brain-damaged patients use their right hand, left-brain-damaged patients their left). Lines of different lengths may be presented, with their center aligned with the mid-sagittal plane of the participant’s body, or displaced leftward or rightward (Daini et al., 2002; Schenkenberg et al., 1980; Vallar et al., 2000). Typically, patients with spatial neglect set the mark shifted toward the end of the line ipsilateral to the side of the lesion (i.e., right-brain-damaged patients with left spatial neglect exhibit a rightward deviation from the geometrical mid-point of the line). Typically, the longer the line, the larger the bisection error (Schenkenberg et al., 1980; Vallar, 2000). The Landmark test (Milner et al., 1993) is a variant of manual line bisection, which does not require a motor response. In the Landmark test, in different trials participants decide which of the two halves of pre-bisected lines is longer or shorter, and communicate their response to the examiner, verbally or by pointing. In patients with left spatial neglect, deciding that the right half of the line is longer or shorter than the left half, independent of their actual lengths, and of the question posed by the examiner (longer or shorter?) indicates a response or output bias. By contrast, judging, erroneously with respect to the actual stimulus, that the left half of the line is shorter (and that the right half is longer) indicates a perceptual underestimation of the lateral extent of the left-sided portion of the pre-bisected line.

3) Drawing: to copy a meaningful or meaningless figure presented in front of the participant (Gainotti et al., 1972) or to draw a familiar figure from memory to a verbal command (Bisiach & Vallar, 2000; Rode et al., 2006). Patients’ performances are evaluated considering the extent of the omissions (complete or partial) of details in both the left- and the right-hand sides of the figure, or of each object, when the drawing includes multiple stimuli (Gainotti et al., 1972). Typically, drawings by patients with left spatial neglect are inaccurate and incomplete in their left-hand side. In copying figures including multiple objects, patients may omit either left-hand side objects, with reference to the mid-sagittal plane of their body showing, as discussed above, egocentric spatial neglect, or the left-hand side of each object, with reference to its coordinate frames, showing allocentric spatial neglect (Walker, 1995).

4) Reading: single words and non-words, sentences and passages of prose. Patients showing neglect dyslexia may commit errors of omission, substitution (Ellis et al., 1987; Kinsbourne & Warrington, 1962) and (less frequently) addition (Siéroff, 2017; Vallar et al., 2010) of letters and words (in reading sentences and passages of prose). Reading errors on words and orthographically legal non-words may be classified as neglect errors, using the “neglect point measure” (Ellis et al., 1987). Neglect errors in reading are defined as “errors in which target and error words are identical to the right of an identifiable neglect point in each word but share no letters in common to the left of the neglect point”. Left neglect dyslexia may be also detected using sentences (Antonucci et al., 1995). More ecologically-based reading material (i.e., passages of prose, as in articles or a menu) has been shown to be sensitive to detect the deficit (Galletta et al., 2014).

Patients with spatial neglect may also show a directional bias, or some other forms of neglect, in Activities of Daily Living (ADL). Prompted by several classical observations (see, for instance, Brain, 1941; Paterson & Zangwill, 1944), assessments based on the simulation of realistic conditions have been proposed to quantify the extent of spatial neglect in ADL. The Behavioral Inattention Test (Halligan et al., 1991; Wilson et al., 1987) is a comprehensive standardized battery that includes both conventional paper-and-pencil and behavioral subtests, in which patients are asked to perform tasks reproducing more practical and everyday life activities. Several batteries, including an ADL scale to evaluate extra-personal and personal spatial neglect have been proposed (McIntosh et al., 2000; Pizzamiglio et al., 1989; Zoccolotti et al., 1992; Zoccolotti & Judica, 1991). The Catherine Bergego Scale (CBS) assesses both the functional deficits in extra-personal and personal spaces (e.g., “forgets to eat food on the left side of his/her plate”; “forgets to clean the left side of his/her mouth after eating”), and the patients’ awareness of them, by comparing the observation by the examiner (the physician, the nurse, a relative) with the patients’ evaluation of their own behavior (Azouvi, 1996, 2002; Azouvi et al., 2003; Bergego et al., 1995). The CBS has been repeatedly reported to be a more sensitive tool to detect spatial neglect than the conventional paper-and-pencil tasks (Azouvi, 2002; Azouvi et al., 2003, 2006). Other ecologically valid and straightforward tasks used to assess spatial neglect in personal space require patients to reach contralateral body parts with their ipsilateral hand (Bisiach et al., 1986), or to explore their own body, using the ipsilateral unaffected arm and hand and, with eyes closed, to remove several Velcro stickers previously attached on their clothes, on both sides of the body (Cocchini & Beschin, 2020).

To summarize, a comprehensive clinical assessment of spatial neglect should include: (i) neuropsychological tests, evaluating both near extra-personal (target cancellation, line bisection, drawing by copy and from memory and reading tasks) and personal (body perception, body awareness, assessed by tests requiring patients to explore the two sides of the body, searching for body parts or objects on them) manifestations of the deficit (see Halligan et al., 2003; Menon & Korner-Bitensky, 2004; Vallar & Bolognini, 2014, for reviews); (ii) ADL scales or ecologic tasks, also assessing the patients’ awareness for the disorder (see for review Azouvi, 2017).

Early Attempts to use Information Technologies (IT) for the Neuropsychological Assessment of Spatial Neglect

With increasing time from stroke onset, many patients recover, and may do well in neuropsychological tests, but still have difficulties in ADLs. A more complex control of spatial attention and awareness, and a major contribution from executive processes, may be required since, in everyday life, every activity is a multiple and not repetitive task, at variance from the previously discussed laboratory tests (Bonato et al., 2010, 2012; Della Sala et al., 2018).

Starting from the late 1990s, several computer- and touch-screen-based tasks have been devised and implemented, to provide a more sensitive and informative assessment, allowing to detect impairments in performance even in patients who do within the normal range in paper-and-pencil tests (Deouell et al., 2005; Erez et al., 2009; Rabuffetti et al., 2002; Schendel & Robertson, 2002).

Computerized tests have indeed some pros, typically recording more precisely information such as accuracy and response latencies throughout the execution of the task. To implement the architecture of the test devised by the experimenter, stimuli may be presented in different locations, simultaneously or as a sequence, with equal or varying time intervals between stimuli, across trials, sessions, or both, and may be repeated many times (Bonato et al., 2010; Buxbaum et al., 2012; Deouell et al., 2005). To manipulate the load posed to the participant’s resources, different levels of difficulty of the task can be easily implemented, varying for instance the duration and the number of the stimuli, and adding a concurrent task (Bonato et al., 2013; Bonato & Deouell, 2013). Additional behavioral measures may also be recorded, such as eye movements (Van der Stigchel & Nijboer, 2010), and data available from touch-screen recording (Rabuffetti et al., 2002), such as number and duration of touches, along with response latency measures recorded throughout the execution of the task (Kim et al., 2010). Furthermore, chances for a ceiling effect (namely, most patients obtain the maximum score in a test) are reduced. Quantitative, continuous measures, appropriate for statistical analyses in single patients, may be easily recorded (Robey et al., 1999), including sensitive individual monitoring of variations in performance levels through repeated assessments. Because of the possibility of easily producing unpredictable settings (e.g., presenting stimuli in random locations of the working space, with different shapes, and time duration), computerized tests can be made harder to learn, making more difficult the development of compensatory strategies on the patient’s part. Computerized tests are thus more suitable for test-retest designs, for instance readily allowing the assessment of the generalization effects of a given rehabilitation treatment (Geusgens et al., 2007). Finally, and importantly, computer-based tasks do not pose a general heavy burden on stroke patients in the post-acute and chronic phases. No adverse reactions, such as headache, fatigue, or boredom, have been reported. Conversely, digital tests have been evaluated as funnier, more precise and safer than the traditional paper-and-pencil ones (see, for instance, Pallavicini et al., 2015; Quinn et al., 2018; Smit et al., 2013; Ulm et al., 2013).

Also Augmented and Virtual Reality-CNADs should address all the above-mentioned issues. Augmented Reality (AR) can be defined as a set of techniques and tools, which add information to physical reality. AR has been used in many fields, including clinical psychology (Chicchi et al., 2015 for review), while its applications to neuropsychology are still limited. Virtual Reality (VR) is a computer-based multisensory, stimulating, and interactive environment, which occurs in real-time. Participants are engaged in activities that appear similar to real-world objects and events (Rizzo et al., 2004). VR can be subdivided in non-immersive (2-dimensional screen presentations, with interaction devices, such as a joystick), and immersive (requiring the integration of computers with further devices, such as Head-Mounted Displays, HDMs, VR controllers, or body-tracking sensors) applications. The immersive applications allow users to experience the virtual environment, and to interact with it, based on the movements of the head or body. For these reasons, starting from the 2000s, several VR paradigms have been implemented in both clinical and research settings (Neguț et al., 2016; Parsons, 2015; Rizzo et al., 2004). As compared to both paper-and-pencil and computer-based assessments, VR may be a useful tool to adequately evaluate ADLs in patients with spatial neglect, through scenarios that resemble everyday life situations (see for reviews Ogourtsova et al., 2017; Pedroli et al., 2015; Tsirlin et al., 2009). Several issues need however to be considered, although several neuropsychological VR paradigms have been developed in the last few years, facilitated by the decreasing costs of the hardware components and the increasing availability of open-access systems. Open issues include: (i) VR-Induced adverse Symptoms and Effects – VRISE (nausea, dizziness, disorientation, fatigue and instability). These symptoms and effects are mainly related to hardware and software inadequacies, not shared/overcome by more contemporary VR hardware and software (Kourtesis et al., 2019a, b). The virtual reality neuroscience questionnaire (VRNQ) is a tool which may facilitate the quantitative assessment of software attributes and intensity of VRISE (Kourtesis et al., 2019b). (ii) Ergonomic features of VR tools. Ergonomic and naturalistic interactions also minimize the risk of VRISE (Kourtesis et al., 2019a). Realistic interfaces, with direct hand interactions and wands with six degrees of freedom of movement, facilitate naturalistic and ergonomic interactions. Direct hand interactions are easier in terms of familiarization with their controls and efficiency and offer a more pleasant user experience; however, they appear less accurate than wands with six degrees of freedom of movement (Figueiredo et al., 2018; Sportillo et al., 2017). Stroke patients often use a wheelchair for locomotion, due to the presence of motor deficits, such as hemiparesis or hemiplegia. VR systems, that can be controlled with one hand (in patients with a hemispheric stroke, the non-paretic ipsilesional upper limb and hand), substantially improve the user experience of the tool itself (Fordell et al., 2011; Kim et al., 2011; Tsirlin et al., 2009). (iii) Psychometric features of the developed systems. Ecological validity is crucial for the assessment of spatial neglect because, with increasing time from stroke onset, patients may still show difficulties in ADLs, despite doing well in neuropsychological paper-and-pencil tests (Bonato et al., 2010, 2012; Della Sala et al., 2018). Immersive VR technology allows to collect data through the employment of dynamic stimuli and interactions with a high degree of control within an ecologically valid environment (Kourtesis et al., 2020; Rizzo et al., 2004). However, immersion depends on the placement, plausibility, and embodiment illusions (Maister et al., 2015; Slater, 2009, 2018). The placement illusion consists in the deception of users to be in a real and not in a virtual environment. The plausibility illusion is the deception of users that the environment reacts to their actions: they thus consider plausible that they are immersed in a real environment. The embodiment illusion refers to the deception that users own the body of the virtual avatar. (iv) Competences. Software developers and researchers are required to have adequate technological abilities, to opt for the appropriate hardware and software and to achieve their research, clinical aims, or both (Kourtesis et al., 2019a). Since a clinical staff typically has no programming skills, cooperation with software developers is needed, to develop and customize these systems (Fordell et al., 2011; Mainetti et al., 2013; Sedda et al., 2013; Sugarman et al., 2011). (v) User experience. This factor is affected by the used hardware, the quality of the sound and graphics, the level of immersion experienced by the user and the level of the enjoyment of the VR experience (Kourtesis et al., 2019b).

Notwithstanding the available evidence, the American Academy of Clinical Neuropsychology (AACN) and the National Academy of Neuropsychology (NAN), raise eight key issues regarding the development, dissemination and implementation of Computerized Neuropsychological Assessment Devices (CNADs), for both research and clinical aims: (i) safety and effectivity, (ii) identity of the end-user, (iii) technical features of hardware and software, (iv) privacy and data security, (v) psychometrics properties, (vi) examinee’s issues, (vii) use of reporting services, and (viii) reliability of responses and results. CNADs could be a standalone device or software, that can be run on a personal computer, a laptop, a tablet, or a smartphone (Bauer et al., 2012; see Kourtesis & MacPherson, 2021 for VR-related issues).

The Review

Up to present, the manifestations of spatial neglect are mainly detected through neuropsychological paper-and-pencil tests, in both clinical and research settings. As noted above, in the subacute and chronic phases post stroke (Rehme et al., 2012), due to spontaneous recovery and cognitive rehabilitation, patients with spatial neglect may do well in neuropsychological tests, but still have difficulty in ADLs (Bonato et al., 2010, 2012; Della Sala et al., 2018). Information Technology (IT)-based devices (computer-, touch-screen- and VR-based) are useful tools for neuropsychological assessment, recording more information than paper-and-pencil tests and allowing to replicate and alter the surrounding world, to evaluate the patients’ behavior in contexts simulating real life (see for reviews Ogourtsova et al., 2017; Pedroli et al., 2015; Tsirlin et al., 2009). However, as previously mentioned, when developing, disseminating and implementing CNADs, for both research and clinical aims, several issues must be taken into account (Bauer et al., 2012; Kourtesis & MacPherson, 2021 for VR-specific issues).

The aim of this scoping review was to present an overview of studies from 2010 onwards, in which IT-based devices were used as a potential tool for the assessment of spatial neglect. Scoping reviews represent indeed a useful instrument to identify and map all available research on a specific topic, and to report information about how studies have been conducted (Munn et al., 2018). Only studies since 2010 onwards were included in this review for two reasons. Firstly, studies prior to 2010 have already been reviewed (Ogourtsova et al., 2017; Pedroli et al., 2015; Tsirlin et al., 2009); secondly, this review aimed at critically verifying whether the tools developed in the last decade successfully target the AACN & NAN Criteria (Bauer et al., 2012; see Kourtesis & MacPherson, 2021 for VR-related issues).

Methods

The norms previously established by the PRISMA-P Group for scoping reviews (Tricco et al., 2018) were adopted. The protocol for this review was not previously registered.

Search Strategy and Selection Criteria

The study search was initially kept broad to capture all relevant articles concerning the use of IT devices [computer, tablet and virtual (VR) or augmented (AR) reality], as possible tools for the clinical assessment of manifestations of the syndrome of spatial neglect. The following electronic databases were used for the identification of papers, up to July 2022: MEDLINE, ScienceDirect, Web of Science and Scopus. The search sting was: (“technology” OR “computer” OR “tablet” OR “virtual reality” OR “augmented reality”) AND [“neglect” OR (“spatial neglect” OR “visual neglect” OR “visuospatial neglect” OR “hemispatial neglect” OR “unilateral spatial neglect”)]. The reference lists of included articles and relevant reviews were checked to identify additional studies. The selection process was performed using Rayyan – a web and mobile app for systematics reviews (access from: https://www.rayyan.ai) (Ouzzani et al., 2016). Titles, abstracts, and full-text articles were screened independently by four out of the six authors (ST, FA, GM and RE), and evaluated for eligibility, based on the following inclusion and exclusion criteria:

  • Inclusion criteria.

    • Related to, and making use of, any of the technologies employed in the clinical assessment of spatial neglect in brain-damaged patients (studies concerning nanotechnologies, technologies applied to eating behavior, nano-molecules, functional Magnetic Resonance Imaging - fMRI, Transcranial Magnetic Stimulation - TMS, transcranial Direct Current Stimulation - tDCS techniques and drug therapies were not included).
    • Focusing on adult brain-damaged patients with spatial neglect (i.e., at least one patient entered the study) and using the abovementioned technologies to assess neglect in patients with a clinical diagnostic aim.
    • Published since 2010.
    • Published in peer-reviewed journals.
    • Published in English.
  • Exclusion Criteria.

    • Technical papers with no clinical application in patients with spatial neglect.
    • Literature reviews.
    • Unpublished data, conference proceedings or articles not reporting quantitative data.

These criteria are in line with those of other published reviews (Bucur & Papagno, 2021, 2022). A preliminary selection based on title, abstract and keywords was run, excluding those articles not matching selection criteria. A further selection was then made by inspecting full text manuscripts and applying inclusion and exclusion criteria. Unresolved papers were discussed by all authors, to reach a consensus. Duplicates were identified and removed through hand search.

In line with several review articles on neuropsychological topics (Bucur & Papagno, 2021, 2022; Ogourtsova et al., 2017; Zhang et al., 2017) we chose: (i) to include only papers published in English, (ii) not to include articles from gray literature (e.g., unpublished Ph.D. thesis) for two main reasons: (a) often, Ph.D. students embargo their dissertations, and (b) good level theses may be subsequently published in journals as research articles (see, for instance, Bucur & Papagno, 2019, 2022). With these inclusion and exclusion criteria, it is likely that only papers, which have passed the peer review process at an international level, are included in the present review.

Results

The literature search retrieved 9764 articles. After duplicate removal, 5344 were screened for further evaluation because they were potentially relevant research articles, that explored the use of technologies (computer, tablet, VR or AR) as possible tools for the clinical assessment of spatial neglect. After Title and Abstract screening, 52 articles were assessed for eligibility, and, finally, 42 articles met all inclusion criteria and were further analyzed. Figure 1 shows the entire search and selection process.

Fig. 1.

Fig. 1

Flow diagram of study selection and inclusion criteria

Articles meeting the inclusion criteria were then categorized into four groups, depending on the digital instruments used to administering the digital tasks: computer (pc and laptop) (13), graphics tablet and touch-screen tablet (8), virtual reality (16), and other (5). The main features of the included studies are summarized in Table 1. In all studies, the presence/absence of spatial neglect was defined according to the neurological exam, a neuropsychological paper-and-pencil assessment, or both. Ten out of 12 studies describing computer-based tools included only patients suffering from a cerebrovascular attack (CVA). Two out these 12 studies (Bonato et al., 2013; Rabuffetti et al., 2012) included patients with different aetiologies (CVA, brain tumor, and traumatic brain injury). One study included also a control group of patients with Mild Cognitive Impairment (Blini et al., 2016). All 13 studies included right-brain-damaged patients; left-brain-damaged patients were included in six out of 13 studies (Blini et al., 2016; Kocanaogullari et al., 2020; Rabuffetti et al., 2012; Villarreal et al., 2020, 2021, 2022). All eight studies, describing graphics tablet- and touch-screen tablet-based assessment tools, included right-brain-damaged patients suffering from CVA; left-brain-damaged patients were included in two out of eight studies (Chung et al., 2016; Pierce et al., 2021). All 16 studies but one (Mesa-Gresa et al., 2011, which included also patients with a brain tumour) using VR-based tools included patients suffering from CVA. Left-brain-damaged patients were included in three out of the 16 studies (Aravind et al., 2015; Aravind & Lamontagne, 2014; Fordell et al., 2011). Immersive VR paradigms were used in all but two studies (Buxbaum et al., 2012; Grattan & Woodbury, 2017). Finally, among the other group, all the five studies included patients suffering from CVA and only two (Peru et al., 2017; Spreij et al., 2020) left-brain-damaged patients.

Table 1.

Characteristics of studies included. Summary of abbreviations and references:

COMPUTER
Report N: Sex (M/F); Age (M ± SD) Aetiology.
duration of disease
Conventional N Assessment General Description Technical aspects User experience Results
Chiba et al. (2010) RBD (N+): 10 (6/4); 70 ± 8.51. RBD: 10 CVA. Drawing task (copy of a daisy)12; Line Bisection20; Line Cancellation test25.

1. Verbal-Line Bisection (VLB) test.

2. VLB plus pointing test.

Horizontal black line presented on computer display. Character string (43 different Japanese language characters) subsequently superimposed on black line. Tasks’ sequence: (a) to identify the subjective midpoint of the line; (b) to read the character nearest to the line mid-point. In VLB the character read by participant is then pointed to by examiner using a cursor; participant required to confirm. In VLB plus pointing, participant points to the character with right index finger (proprioceptive feedback) and communicate if it is the one just read.

Stimuli presented on laptop computer, equipped with 39-cm liquid-crystal display. Windows operating system. Microsoft PowerPointⓇ software to show images. Computer placed on a table in front of participant, with centre aligned with participant’s body’s mid-sagittal plane (viewing distance about 45 cm). - In both the VLB task (8 out of 10 patients, 80%), and the VLB task plus Pointing Task (9 out of 10 patients, 90%), rightward deviation from the objective mid-point of stimuli.
Rabuffetti et al. (2012)

RBD (N+): 55 (32/23); 64.72 ± 14.5.

RBD (N-): 66 (37/29); 56.1 ± 15.8.

LBD (N-): 72 (41/31); 60.6 ± 14.2.

HC: 119 (33/86); 58.92 ± 15.2.

RBD (N+): 48 CVA, 3 TBI, 3 BT, 1 Missing; 28 (1–3 m), 12 (4–12 m), 13 (> 12 m), 2 (missing).

RBD (N-): 44 CVA, 11 TBI, 6 BT, 5 Missing; 24 (1–3 m), 12 (4–12 m), 22 (> 12 m), 8 (missing).

LBD (N-): 56 CVA, 9 TBI, 3 BT, 4 Missing; 26 (1-3 m), 12 (4–12 m), 25 (> 12 m), 9 (missing).

BIT3’s Line Bisection; Letter Cancellation18; Line Cancellation24; Sentence Reading32; Wundt-Jastrow Illusion test44.

Visual stimuli (120: 40 targets, 80 distractors, letters, or shapes),

presented on touch-screen monitor. Participants’ task: to touch with the index finger all targets. Sliding behaviour precluded by real-time testing software, allowing a touch-and-go behaviour. Maximum test duration: 10 min.

19” touch-screen monitor. Participants seated in front of screen with trunk’s mid-sagittal plane aligned with screen centre. - RBD N + patients slower and less efficient in search than HC, with more leftward and less downward direction of exploration. RBD N- patients: same pattern, halfway between RBD N + patients and HC.
Bonato et al. (2013) RBD (N-, according to paper-and-pencil assessment): 10 (5/5); 67. RBD: 9 CVA and 1 BT; 92 d. BIT3 Line, Letter and Star Cancellation subtests. Participants detect briefly presented lateralized targets in Single Task (ST) and Dual Task (DT) conditions. DT conditions: (i) Visual Dual Task (VDT, to report the identity of a centrally presented letter before reporting target position), (ii) Auditory Dual Task (ADT, counting by twos twice from a heard number, before reporting target position). Participants seated at a 60-cm distance from a 15-in. computer monitor. Tasks programmed and administered using E-Prime (Bonato et al., 2010). -

Computer-based tasks detect deficits in contralesional side of space in RBD patients, classified as N- by paper-and-pencil cancellation tests.

Deficits more severe under DT (VDT and ADT) conditions: about 70% of contralesional targets not detected.

Ulm et al. (2013)

RBD (N+): 10 (6/4); 60 ± 8;

HC: 10 (4/6); 68 ± 9.8.

RBD: 10 CVA. Of them, 5 (subacute group): 1.7 ± 0.8 m; 5 (chronic group): 20 ± 20.0. CAV’s Computer-Based Extinction Test and examination of the visual field9; NET28’s Star Cancellation, Line Bisection, Figure Copying, and Clock Drawing subtests; DSS for visual and tactile extinction.

Circle-Monitor (CM) tests.

1. Star Cancellation test.

2. Line Bisection test.

3. Dice Task (DT).

4. Puzzle Test (PT).

Star Cancellation and Line Bisection tests based on the same tasks in the NET28; DT, a simplified version of Baking Tray Task2; PT based on Hooper Visual Organization Test (HVOT)17.

Participants completed all tasks using their right hand to reach the touch-screens, while sitting inside the CM. Total duration: about 10–15 min.

CM: eight touch-screens arranged in a circle and connected to computer outside the CM. Participants seated in the middle of the CM, at a distance adequate to reach monitors comfortably. Participants’ behaviour also supervised by a video camera. CM’s tasks rated as more complex and fun, than paper-and-pencil versions, by N + patients and HC. RDB patients’ performances at CM’s Star Cancellation and Line Bisection test comparable to those obtained in paper-and-pencil versions. More detailed analysis of time course of spatial exploration and additional scores (Crossing and Latency indexes) provided by CM. DT and PT subtests more sensitive to detect neglect than paper-and-pencil tests.
Jee et al. (2015)

RBD (N+): 11 (10/1): 63.54 ± 14.72.

HC:

-Group #1: 42 (21/22): 24.65 ± 2.87.

-Group #2: 11 (5/5): 23.6 ± 2.18.

RBD (N+): 11 CVA, 6 m. -

Semi-computerized Line Bisection test (LBT).

Twenty lines and dotted micro-patterns pre-printed with black ink on 21.5 × 28cm papers. Lines randomly selected and presented to participants with the instruction to cut each line in two equal halves, by placing a small pen notch through it, as closest to its centre as possible. Computed scores: (i) percent deviation, (ii) assessment duration (sec), (iii) neglected lines. Scoring: (i) LBT data sent to computer for computation and scores displayed on monitor, (ii) hand-drawn results given to two blinded raters, for independent scoring.

System installed on computer: position pattern recognizing the electronic pen, micro-pattern printed paper, installable LBT, assessment program. Written or drawn information detected and sent by the pen from a position-coded micro-pattern printed paper to computer, with LBT program installed. - Feasibility assessment of e-system first performed in HC, to evaluate intra- and inter-rater reliability. e-system then used in patients. Reliability of automatic scoring suggested by high inter-rater correlation between data scoring computed by automatic and manual procedures, the latter performed by two independent assessors.
Blini et al. (2016)

LBD (N-, according to neurological and neuropsychological exams): 10 (6/4); 53.2 ± 11.7.

HC: 10; 65.8 ± 8.52.

MCI: 8 (5/3); 69 ± 11.61.

LBD (N-): CVA, 456.1 ± 773.11 d. BIT3’s conventional tests2. DSS paradigm 11 to assess visual half-field deficits and extinction. Patients asked to report position of target(s), shown unilaterally or bilaterally on a computer monitor. ST and two DT conditions (visual vs. auditory). Target: a white disk, presented against a black background for 100 ms. Simultaneously with the lateralized target(s), visual shape shown at fixation point, or sound presented binaurally through earphones. ST condition: to report target(s) position. DT conditions: report both target position and visual central shape or presented sound. Patients individually tested in quiet room, sitting at about 60-cm from 19-in computer monitor. - No spatial deficits in baseline neuropsychological assessment. Defective report of contralesional stimuli, both single and under conditions of DSS in the DT condition. No spatial bias in HC (performance at ceiling), and in MCI patients.
Mizuno et al. (2016)

RBD (N+): 2 (1/1); 62.5.

RBD (N-): 1 M; 73.

HC: 16 (11/5); 57.3 ± 10.7.

RBD (N+): 2 CVA, 7 w.

RBD (N-): 1 CVA, 4 w.

BIT3’s conventional subtests.

Four visual exploration tests presented on a touch-screen monitor.

1. Star Cancellation test: cancelling targets by touching them.

2. Circle test: selecting complete circles, among incomplete ones.

3. Visuomotor Search test: to touch the screen where a Star target shows up in a random position.

4. Visual Search test: to push a button when a Star target shows up in a random position.

Participants seated at a desk in front of a 32-in. liquid crystal touchs-creen display, placed at a 45-cm distance from their eyes. - In Star and Circle tests, RBD N + patients slower than RBD N- patient and HC.
Machner et al. (2018)

RBD (Severe N+): 12 (9/3); 69 ± 3.

RBD (Moderate N+): 12 (7/5); 68 ± 3.

RBD (N-): 10 (3/7); 71 ± 3.

HC: 11 (3/8); 69 ± 4.

RBD (Severe N+): 12 CVA, 5 ± 1 d.

RBD (Moderate N+): 12 CVA, 4 ± 1 d.

RBD (N-): 10 CVA, 4 ± 1 d.

Bells Cancellation test5; BIT3’s (German version) Line Bisection, Star Cancellation and Text Reading; CBS7; Figure Copying test13. CoC8 computed for Cancellation tests.

1. Desk task. Desk scene images (30 objects each, 100 trials) shown on 24-in. monitor. Participant’s task: to search for a paperclip (blue or red), to press a response button when the target was found, and to name its colour. Recorded data: detection rate, search duration and eye movements

2. Posner Reaction Time (RT) task. Computerized variant of Posner cueing paradigm (18.4-in. monitor). Data: detection rate and latencies.

1. Desk task. Participants seated at about 60-cm in front of a 24-in. widescreen thin-film transistor monitor: contact-free remote eye tracker running at 50 Hz.

2. Posner RT task. Notebook with a 18.4-in. thin-film transistor widescreen monitor.

- Most RBD patients (also those classified as N- according to CBS and performing at ceiling at paper-and-pencil tests) show ipsilesional bias in Desk Test: fewer contralesional targets detected, and ipsilesional deviation of exploratory eye movements. High correlation of these rightward biased indexes with decreased detection rate and increased RTs for left-sided contralesional targets in Posner RT task and neglect-related functional impairment scores.
Rossa et al. (2019)

RBD (N+): 4 (1/3); 55.5 ± 17.2.

HC: 8 (5/3); 49.37 ± 12.97.

Unavailable. Bells Cancellation test5.

READAPT

Bells Cancellation test: computerized version.

Visual scenes made of 49 items (1 target, 48 distractors). In each trial participants communicate the target’s presence/absence, using the left/right mouse button. Three different versions: (i) Standard (one block, 80 trials). (ii) Items (five blocks, 14 trials each. Position of distractors on horizontal plane changed in every block). (iii) Interference (five blocks, 14 trials each. Percent degradation of the scene on the horizontal plane changed in every block). Eye movements recorded by eye tracker.

REDAPT. Web application: a set of programs written in PHP and JavaScript, coupled with eye-tracking recorder.

Participants seated in front of a computer in the straightest position possible, at a 60-cm distance from the screen. Eye tracker placed under the screen at a 30-45-cm distance from participants.

- More target stimuli (total score and left-sided targets, not right-sided targets) omitted, and more time to complete the test by RBD left N + patients, compared to HC.
Kocanaogullari et al. (2020)

RBD (N+): 4 (3/1) 49 ± 10.96.

LBD (N+): 1 (F) 57.

RBD (N-): 1 (F) 68.

LBD (N -): 4 (3/1) 68 ± 9.48.

BILATERAL (N-): 1 (F) 69;

RBD (N+): 4 CVA, 42.33 ± 63.00 d.

LBD (N+): 1 CVA, 7d.

RBD (N-): 1 CVA, 17d.

LBD (N -): 4 CVA, 10.2 ± 6.30 d.

BILATERAL (N-): 1 CVA, 10 d.

BIT3 Starry Night Test. To press a key on a keyboard when detecting a visual target (a red dot) among distractors (smaller green dots), on a computer screen monitor. Response times recorded. EEG recording throughout the test. Computer screen monitor with 17.23° by 9.74° viewing area. EEG recorded through 16 electrodes located at Fp1, Fp2, F3, F4, Fz, Fc1, Fc2, Cz, P1, P2, C1, C2, Cp3, Cp4, P5 and P6, according to 10–20 system with 256 Hz sampling frequency. -

EEG data related to slower response to targets by patients. Three most informative EEG channels for detecting neglect: Cz, P1, and F3 (broadly corresponding to regions in the anterior frontal cortex and the superior posterior parietal cortex).

Response times not analyzed.

Villarreal et al. (2020)

RBD: 20 (9/11); 53 ± 8.

LBD: 20 (15/5); 51 ± 9.

HC: 20 (8/12); 46 ± 15.

RBD: CVA, 106 ± 45 d.

LBD: CVA, 105 ± 45 d.

Bells Cancellation test5.

Two computerized DTs

1. Detect task. (i) To respond, by pressing a mouse button, to a red sphere flashing among other coloured ones in the peripheral visual field task. (ii) To respond when the digit “2” appears among different numbers presented in the centre of the screen.

2. Crash task. (i) To respond to a visual target (the collision of two moving grey spheres) in the peripheral visual field task (ii) To respond verbally when a number presented in the centre of the screen is twice as high as the immediately preceding one.

In both peripheral and central vision, targets never presented simultaneously. Scores: correct responses, omission errors, RTs.

Active Space, based on near-field imaging technology. Several means of generating visual stimuli and measuring RTs. Main generator of visual stimuli: short-throw video projector producing a display 173-cm high and 277-cm wide (1.9 mm pixel size). Screen midpoint located 120-cm high from the floor. - More left-sided targets missed by RBD patients than by HC. No neglect in RBD, assessed by paper-and-pencil Bell Cancellation test. No difference in target omission between LBD and HC. No differences in RTs in both Detect and Crash tasks among RBD, LBD and HC.
Villarreal et al. (2021)

RBD: 20 (9/11); 53 ± 8.

LBD: 20 (15/5); 51 ± 9.

HC: 20 (8/12); 46 ± 15.

RBD: CVA, 106 ± 45 d.

LBD: CVA, 105 ± 45 d.

Bells Cancellation test5.

Computer-based

1. Lateralized target computer task. Participants in front of a computer monitor at 60-cm distance. Three different stimuli presented concurrently: visual black dot target, visual central letter, and spoken number. Dot target shown in the left, or right hand-side of the screen, or bilaterally. Participants’ task: to report the position of the dot target(s), disregarding the central letter and the spoken number.

2. Ball Rain Task. To look at coloured spheres continuously appearing from the top of the screen in a fast downward motion. To respond, pressing the mouse button, only to red spheres.

15-inch computer monitor running E-Prime software. Video projector to generate visual stimuli, producing a 173 × 277-cm display on the wall (pixel size 1.9 × 1.9 mm). Task application implemented using LabVIEW™ systems engineering software. -

1. Lateralized target computer task. With bilateral stimuli, more left-sided than right-sided target missed by RBD patients, than by LBD patients and HC.

2. Ball Rain Task. More targets contralateral to the side of the lesion missed by RBD and LBD patients than by HC. Higher RTs in patients. No spatial biases in Bells Cancellation test.

Villarreal et al. (2022)

RBD: 20 (9/11); 53 ± 8.

LBD: 20 (15/5); 51 ± 9.

HC: 20 (8/12); 46 ± 15.

RBD: CVA, 106 ± 45 d.

LBD: CVA, 105 ± 45 d.

Bells Cancellation test5.

1. Large-screen version of the Twinkle task. 180 five-pointed star-shaped black outline figures arranged on a white background. 110 intact figures (distractors) and 70 figures with one point missing (target). Participants used a mouse with their dominant hand to move an on-screen cursor. Targets selected by pressing the left button of the mouse.

2. Finnish version of the Visual and Auditory dual-tasks developed by Bonato et al. (2010, 2013) (see lines above for task descriptions).

Paper-and-pencil version of the Twinkle Task consisting of a screenshot of the large-screen version, with identical stimuli and proportions, presented on an A4 size paper.

Active Space was used to generate a large-screen version of the Twinkle Task. Main generator of visual stimuli: short-throw video projector producing a display 173-cm high and 277-cm wide.

(see Villarreal et al. 2020).

For technical aspects about the Visual and Auditory dual-tasks see General Description by Bonato et al. (2013)

-

Neither the computer nor the paper-and-pencil version of the Twinkle task was able to show spatial bias in RBD patients.

In the Visual dual-task RBD patients missed significantly more left-sided targets than HC in both unilateral and bilateral trials. They also missed significantly more left-sided than right-sided targets only in the bilateral trials of the Auditory dual-task.

GRAPHICS TABLET AND TOUCH-SCREEN TABLET
Report N: Sex (M/F); Age (M ± SD) Aetiology; duration of disease Conventional N Assessment General Description Technical aspects User experience Results
Liang et al. (2010)

RBD (N+): 18.

RBD (N-): 36.

(Whole sample composed as follows: 30 M/24F; mean age: 75 ± 7.4.).

RBD (N+): 18 CVA.

RBD (N-): 36 CVA.

BIT3.

Digitized tests

1. Target Cancellation (line, O/X).

2. Figure Copying (cross, cube, star, square).

3. Figure Completion (diamond, man, house). Participants, presented with half of a schematic figure split vertically, complete it.

4. Drawing from Memory (square, cube).

Test overlays placed and secured individually on Wacom UD1212 (A4) graphics tablet. Drawings executed using a cordless inking pen providing the same writing experience and user feedback as felt in a non-automated test procedure. - Performance of RBD patients, both N + and N-, at digital tests correlated with BIT scores. Overall efficiency of the testing process improved by the digital procedure: more indexes, compared to those of paper-and-pencil tests, extracted; scores automatically computed.
Smit et al. (2013)

RBD (N+): 31.

RBD (N-): 2.

(Whole sample: 19 M/14F; mean age: 60.36 ± 13.30).

RBD: 33 CVA, 63.73 ± 37.74 d. -

Digitized tests performed on a tablet surface using a digital stylus.

1. Object Cancellation.

2. Letter Cancellation (scores: total number of omissions, total time for test completion, search time in ipsilesional and contralesional sides of display, horizontal and vertical CoC).

3. Line Bisection (distance of mid-point set by participant from physical mid-point; total time of execution).

22-inch interactive WACOM (PL2200) tablet screen. DiagnoseIS for programming. Tablet driven by laptop, to monitor stimuli and patients’ performance on the experimenter’s laptop. - Patients able to deal with the digital instrument, to comply with instructions, performing and completing the proposed tests within the allocated time frame indicating feasibility of computer-based assessment in CVA patients.
Pallavicini et al. (2015)

RBD (N+): 5/3.

RBD (N-): 7/1.

(Whole sample’s mean age: 66.1 ± 11.9).

RBD (N + and N-): 16 CVA, 20.31 ± 19.28. BIT3’s Line and Star Cancellation subtests; Card-Dealing task from Semi-Structured Scales for the Functional Evaluation of Hemi-Inattention29.

NeglectApp

1. Two target cancellation tasks (with/without distractors corresponding, respectively, to the paper-and-pencil Line and Star Cancellation tests).

2. Card-dealing task, recreated in a Virtual Environment (VE), shown on the tablet.

NeglectApp tests carried out with a stylus for touchscreen on an iPad2 (screen size active area: 47.70 × 26.82-cm). System Usability Scale (SUS)38. High usability of NeglectApp, as shown by the mean score of the whole sample.

Computer based Line and Star Cancellation tests as effective as paper-and-pencil versions in detecting left neglect.

Card-Dealing task more sensitive than functional scales to detect neglect.

Vaes et al. (2015)

RBD (N+): 20 (12/8); 61.65 ± 9.85.

HC: 20 (9/11); 61.45 ± 9.37.

RBD (N+): 20 CVA. -

Digital Visuospatial Neglect Test Battery, nine tasks.

- Two Cancellation tests: digitized versions of the Bells Cancellation test5 and of the Diamond Cancellation test.

- Two Line Bisection tests: a digitized version of Schenkenberg et al.’s (1980) Line Bisection test14 and a coloured Rectangle Bisection test.

- Visuospatial Search task.

- Drawing by copy task: alarm clock, complex butterfly.

- Visual DSS test.

- One spatial memory task: remembering 12 pictural stimuli placed to the left (six) and to the right of the display.

- Visuospatial navigation test.

After task execution, data processed online.

Metrisquare DiagnoseIS (MDIS) software package used to design stimulus pages, tasks, and the matching score system. A DTU-2231 Wacom (active area: 47.70 × 26.82-cm), with dual-screen technology. Participants used a pen at the display surface. Experimenters observed and reported the results at their computer screen. - Performance difference between left N + patients and HC, with patients performing worse for almost all indexes extracted by the digital apparatus, with a rightward bias: CoC in target Cancellation tasks; Centre of Drawing CoD) in the Drawing test; Centre of Navigation (CoN) in the visuo-spatial navigation test.
Chung et al. (2016)

RBD (N+): 20 (10/10); 69.2 ± 10.2.

RBD (N-): 10 (2/8); 65.5 ± 13.0.

LBD (N-): 10 (5/5); 61.1 ± 12.0.

HC: 10 (5/5); 46.0 ± 15.2.

RBD (N+): 20 CVA, 8.0 ± 6.3 d.

RBD (N-): 10 CVA, 4.0 ± 1.8 d.

LBD (N-): 10 CVA, 6.8 ± 6.6 d.

Line Bisection test23; Star Cancellation test37.

Total Neglect Score calculated from scores of the two tests.

Computer Table Setting Test (CTST). A virtual table with 12 dishes shown in the middle of the screen. Participants set the table by dragging individual plates, located below the table, to the top of the table. Three parameters extracted: horizontal deviation from the midline, selection tendency for dishes on the right, elapsed time to set the table. CTST developed to run on iPad. -

Overall sensitivity of the test higher than that of paper-and-pencil (Line Bisection, Star Cancellation) tests.

RBD N + and N- patients discriminated by scores in the horizontal deviation parameter.

Willer et al. (2016)

Experiment #1:

Patients: 54 (34/20); 63.8.

HC: 48 (22/26); 67.5.

30 patients and 11 HC from the sample above were then re-tested.

Experiment #2

Patients: 16 (11/5); 67.3.

Experiment #1:

Patients: 54 CVA, 17 d.

Experiment #2:

Patients: 16 CVA, 48.4 d.

BIT3’s Line, Star and Letter Cancellation subtests; DKEFS10’s Color-Word Interference test; WAIS-IV41’s Digit Span and Coding subtests; Western Aphasia Battery (WAB)43 (part II b and IV a); WMS-IV42’s Symbol Span and Designs subtests.

Cognitive Assessment at Bedside for iPad (CABPad). Digitized versions of: Rating of Anosognosia, Motor Speed for Hands, Speech Comprehension, Picture Naming, Verbal Fluency (phonemic and semantic), Timed Neglect Test, Baking Tray Test, Attention Span (symbols), Working Memory (symbols), Arrow Stroop, Episodic Memory (pattern locations), Symbol Digit Coding, GDS16 short form.

Phase 1: patients and HC tested twice with a one-month interval

Phase 2: patients tested with CABPad and conventional tests for neglect.

CABPad, developed by Palle M. Pedersen. Made available in Apple’s App Store. - Medium-to-high correlations between almost all CABPad’s subtests and paper-and-pencil tests. Performance of left N + patients worse than that of HC in Timed Neglect, but not in Baking Tray subtest.
Quinn et al. (2018)

Sample: 48 patients (28/20); 63.

(3 N + according to gold-standard assessment; no more data available for the side of lesion site).

Sample: 48 CVA, 38 d.

Object Bisection test. Six horizontal lines, drawn offset from the center.

Goldman or Octopus perimetry’s Visual Field assessment. Line Cancellation test24.

StrokeVision app.

Assessment of sensory (visual acuity, visual field), and spatial and exploratory (line bisection, target cancellation) processes.

App used with a Google Nexus 10 tablet device. 0–10 scale (0 = completely unacceptable, 10 = perfect acceptability), administered to both clinicians and patients. Feasibility and acceptability for stroke patients: app-based assessment completed by patients; modality of testing found acceptable.
Pierce et al. (2021)

RDB (N+): 12 (10/2) 62 ± 11.33.

RDB (N-): 17 (12/5) 60.06 ± 8.8;1.

LBD (N-): 10 (5/5) 59.7 ± 12.26.

HC: 14 (6/8) 66.14 ± 7.87.

RDB (N+): 12 CVA, 9.47 ± 5.79 m.

RBD (N-): 17 CVA, 11.03 ± 8.69 m.

LBD (N-): 10 CVA, 4.65 ± 1.94 m.

Bells Cancellation test5; Line Bisection test22; Apples test1.

Performance scored as the difference in target omissions on the left vs. right side of the sheet. CoC8 measure also scored.

Manual exploration task.

Digitized task, running on a tablet, assessing manual spatial exploration. Participants, blindfolded, tap on the screen using their index finger until the target (a rectangular area 1/8 by 1/5 of the screen size that shows up in a random location on each trial) is found. When the target is tapped, a chime sound is played, and the target moves to a new location. Duration of each trial: 40 s.

Touch screen tablet computer (Microsoft® Surface Pro 4; 26 × 17-cm screen area; 2736 × 1824 pixel resolution) running a custom MATLAB® (Mathworks, Natick, MA) script. - Four out of 12 RBD patients, classified as N + by paper-and-pencil tests, show a significant rightward shift of the average horizontal position in the Manual exploration task, compared to both N- patients and HC. In six patients the rightward shift differs only from the shift of HC. No correlation between shift in the Manual Exploration Task and errors in the paper-and-pencil Bells and Apples Cancellation tasks and shift in Line Bisection.
VIRTUAL REALITY-BASED ASSESSMENT
Report N: Sex (M/F); Age (M ± SD)

Aetiology;

duration of disease

Conventional N Assessment General Description Technical aspects User experience Results
Kim et al. (2010)

RBD (N+): 16 (10/6); 52.9 ± 16.8.

RBD (N-): 16 (11/5); 60.1 ± 12.1.

RBD (N+): 16 CVA.

RBD (N-): 16 CVA.

Letter Cancellation test19; Line Bisection test22. Three-Dimensional Virtual Street Crossing Program Contents. Participants, immersed in the VE, required firstly to place in front of them an avatar presented in the centre of the screen; then, asked to make the avatar safely cross the street by stopping a virtual car that could approach the midline at four different velocities; participants could receive visual or auditory cue. 16 missions performed. A personal computer system with a 3D graphics acceleration card and speakers used to create the VE. Participants immersed in the VE by means of a head-mounted display (HMD). Participants’ head movements measured using a head tracking system with 3 deg. of freedom in the VE. - Significant rightward deviation from the midline in locating the avatar by RBD left N + patients, compared to RBD N-patients. Percent deviation in Line Bisection test correlates with deviation from midline in the VR-based task. RTs of N + patients higher when the virtual car is approaching from the left hand-side of the screen; failure rate (i.e., virtual car not detected by the time the avatar’s position is reached, despite cues), higher than those of RBD N- patients.
Tanaka (2010) RBD (N+): 9 68.7 ± 7.61. RBD (N+): 9 CVA, 4–27 w. BIT4’s (Japanese version) Line and Star Cancellation subtests; modified version of Halligan et al.’s (1991) checklist of everyday neglect behaviours. Paper-and-pencil Line and Star Cancellation tests replicated in a VE. Paper tests photographed by a digital camera. Images altered, by increasing (zoom-in condition) or reducing (zoom-out), their size and displayed in the HMD. Digital camera, HMD and digital video camera, recording participants’ head movement as qualitative motion analysis. - VR cancellation tests: lower scores in the left-sided than in the right-sided halves of the VR working sheet area in all RBD left N + patients. Only in three out of 9 patients, scores lower than the cut-off in paper-and-pencil Line and Star Cancellation tests.
Fordell et al. (2011)

N+: RBD: 8; LBD: 1. 6/3; 73.3 ± 12.

N-: RDB: 11; LDB: 11. 16/6; 74.4 ± 10.8.

N+: 9 CVA.

N-: 22 CVA.

(Duration of disease range for the whole sample: 1–20 w).

Baking Tray Task2; BIT3; Line Bisection21; Star Cancellation test36; Visual Extinction test40.

VR-DiSTRO battery: four subtests, developed into an interactive and immersive 3D experience, evaluating neglect by: (i) Baking Tray Task (VR-BTT); (ii) Star Cancellation Task (VR-SCT); (iii) Line Bisection (VR-LB); (iv) Visual Extinction (VR-EXT).

Total time for the assessment: about 15 min.

Desktop computer with stereo-capable graphics card, stereo headphones, robotic pen, separate numeric keyboard, 19” CRT monitor, shutter glasses for stereoscopic vision.

Glasses, monitor, and software create a virtual 3D world. Using the pen, the patient’s hand perceives “virtual” touch sensation and realistic force feedback.

Overall, satisfaction in the experience, and full comprehension of instructions, reported by patients. No reported adverse effects. In all tasks, rightward spatial exploration bias in RBD N + patients. For VR-DiSTRO, very high values of sensitivity and specificity for detection of neglect. For performances in Baking Tray and Visual Extinction tests the highest correlation with BIT total score, and high sensitivity and specificity. Lower sensitivity for Star Cancellation and Line Bisection tests of the VR-DiSTRO.
Mesa-Gresa et al. (2011)

Patients: 25 (14/11); 51.2 ± 12.62.

(According to BIT scores, N+: 5; N-: 20).

Patients: 22 CVA and 3 BT (L or R), 505.42 ± 335.11 d. BIT3; Colour Trail Making Test (CTT); Conners’ Continuous Performance Test (CTT-II). VR Street Crossing Test (VRSCT). Task: to cross a virtual two-way road to arrive at a destination point and then to come back to the starting point. When an accident occurs, patients receive emotional audio-visual feedback and new instructions. Session completed when patients perform two complete routes with four or less accidents. Measures: number of times participants look leftwards and rightwards; total time needed to complete the task; total number of accidents; completion/non-completion of the task.

Conventional panoramic 47” LCD monitor and 5.1 surround sound system.

Interaction with and navigation in VE, by conventional joystick. Optical system to track patient’s head movements.

- In RBD left N + patients higher number of accidents than in RBD N- patients, with no left vs. right differences. No correlation between VRSCT and N + patients’ BIT scores.
Peskine et al. (2011)

RBD patients (7 N+): 9 (5/4); 50 ± 15.

HC: 9 (5/4); 50.6 ± 16.1.

RBD patients: 9 CVA; 16.11 ± 30.16 m. Bells Cancellation test6; CBS7. Participants immersed in a VR town, in which they can move forward via a mouse click. Patients, sitting in a swivel chair, have to turn on their vertical axis, to change their point of view and moving direction. Task: locating the primary target (swings in a park) and counting bus stops (6 on one side and 7 on the other side of the street), along the way. Participants equipped with HMD. coupled with electromagnetic sensor system. - More targets and more left-sided bus stops omitted by RBD patients than by HC. Convergent results between VR-based tasks and paper-and-pencil tests only in 3 out of 12 patients.
Sugarman et al. (2011) RBD (N+): 1 (F); 66. RBD (N+): CVA; 15 m. Paper-and-pencil Cancellations tasks (no details available about specific tests). VR React task of the SeeMe System: virtual balls appear randomly on both sides of the screen. Task: to touch the virtual ball within a set amount of time. Number of missed balls, and movement time to reach the target, recorded by the system.

SeeMe: projected video capture, VR system with algorithms for movement and position recognition and analysis.

Single screen-mounted camera, and vision-based tracking system, to capture and convert participant’s movements for processing.

Short Feedback Questionnaire (SFQ)34. Patients gave a full score of 5 out of 5 for the items of “enjoy”, “feeling of control” and “success in the virtual tasks”. Performance of RBD patient in paper-and-pencil tests within normal range. In VR React task half of left-sided and none of right-sided targets missed. Patient’s average movement time to targets presented in left side of screen higher than that to right-sided targets.
Buxbaum et al. (2012)

RBD patients: 70 (39/31); 59.5 ± 10.6.

HC: 10 (5/5); 62.2 ± 15.1.

RBD patients: 70 CVA, 29.2 ± 23.7 m. Modified version of Fluff test14; timed version of Bells Cancellation test5; BIT3’s Letter Cancellation and Line Bisection subtests; Buxbaum et al.’s (2004) Laser Line-Bisection task; modified version of Buxbaum et al.’s (2008) Moss Real World Navigation (RWN) test. Virtual Reality Lateralized Attention Test (VRLAT) (Buxbaum et al., 2008; Dawson et al., 2008): participants travel along a virtual, nonbranching path, either propelling themselves using a computer joystick (participant condition), or passively viewing the environment while an examiner navigates the path at a constant rate (examiner condition). Participants asked to identify virtual objects (coloured trees and statues of animals) on either side of the path, and to avoid colliding with them. Array conditions: (i) simple (contains only target objects); (ii) complex (contains targets and distractors); (iii) enhanced (as the complex condition, with the addition of auditory and small visual moving distractors). VRLAT run on a PC. Participants seated 34 in. away from a 15.5 by 27.5-in. flat-screen video display, interfaced with the VRLAT via a Logitech® Attack 3 joystick. Software on the Unreal Engine 2 programmed by Digital Media Works.

HC and 64 patients completed all six conditions of the VRLAT.

Six patients unable to complete the participant-driven conditions, due to difficulty in operating the joystick.

Based on performances on paper-and-pencil tests, 32 (50%) out of 64 patients classified as left N+. This percentage increases to 36 (56.3%) using the VRLAT.

In the Moss Real World Navigation test eight RBD patients make left-sided collisions. All eight patients classified by VRLAT as N+; only five out of eight patients classified as N + by paper-and-pencil tests.

Dvorkin et al. (2012)

Experiment #1:

RBD (N+): 8 (4/4); 59 ± 12.20.

RBD (N-): 9 (8/1); 59.6 ± 7.61.

HC: 9 (5/4); 55.7 ± 8.

Experiment #2:

RBD (N+): 3 (2/1); 50.33 ± 11.23.

HC: 2.

Experiment #1:

RBD (N+): 8 CVA, 18.62 ± 20.87 m.

RBD (N-): 9 CVA, 64.11 ± 41.40 m.

Experiment #2:

RBD (N+): 3 CVA; 20.33 ± 15.04.

BIT3’s conventional subtests.

Virtual Environment for Spatial Neglect Assessment (VESNA). Experiment #1: on each trial, participants instructed to press a response button, when a target is detected. 105 static targets distributed on a Cartesian grid, from right to left of body mid-line, above, below, and at eye level, and in peri-personal and far extra-personal space.

Experiment #2: 35 targets distributed on a polar grid, a fan-like pattern of targets centred between the participant’s eyes, to assess whether spatial deficits are influenced by the rectangular arrangement of targets.

Images displayed by cinema-quality digital projector. Over 5-foot-wide 1280 × 1024 pixel image resulting in a 110° wide viewing angle. Head motion tracked by sensors attached to stereo shutter glasses worn by participants, sitting in a dark room on a chair placed in front of system and holding a response button in their right arm. Head position kept firm by chin rest. -

All RBD patients classified as N + by VR paradigm.

Only 50% of patients classified as N + by BIT. In the VESNA, HC detect all targets. In left N + patients higher RTs and lower detection rate for left- than for right-sided targets, and for near than for far targets.

Aravind & Lamontagne (2014)

RBD (N+): 10 (3/7); 59.9 ± 8.55.

LBD (N+): 2 (1/1); 64.5 ± 10.60.

RBD and LBD (N+): 12 CVA; 13.56 ± 24,28 m. Bells Cancellation test5; Line bisection test22.

Virtual Environment. A room with dimensions matching those of a physical room. A blue circular target presented on the wall at the far end of the virtual room; three red cylinders (obstacles) placed in front of a possible point of collision.

1. Locomotor task. Patients to walk towards the blue target, avoiding collisions with approaching moving obstacles, coming from the centre, left- or right-hand sides of VE. Control trials devoid of any moving obstacle.

2. Perceptuo-motor task. Patients seated, responding using a joystick. Task: to press the joystick button when an obstacle is detected, and not to respond in the absence of obstacles.

Locomotor and Perceptuo-motor tasks performed while patients view the Virtual Environment by nVisorSX60 HMD. Positions of three reflective markers placed on HMD, tracked by 12-camera Vicon-512™motion capture system and fed to CAREN 3™ VR software, to provide the real-time update of the patients’ perceived position and orientation in the Virtual Environment. -

1. Locomotor task. Five out of 12 patients collide with contralesional, eight with middle and none with ipsilesional obstacles. Patients’ performances unrelated to clinical measures of neglect (Bells Cancellation, Line Bisection tests). Fewer collisions in slower walkers.

2. Perceptuo-motor task. Longer latencies in patients to detect contralesional vs. ipsilesional obstacles.

Aravind et al. (2015)

RBD (N+): 10 (3/7); 59.9 ± 8.55.

LBD (N+): 2 (1/1); 64.5 ± 10.60.

RBD and LBD (N+): 12 CVA; 13.56 ± 24,28 m. Bells Cancellation test5; Line Bisection test22; Motor Free Visual Perceptual Test (MVPT)26.

Virtual Environment as in Aravind & Lamontagne (2014). Tasks:

1. Obstacle Detection Task. Patients, by clicking a joystick button, respond on perceiving the onset of obstacle motion).

2. Locomotor Obstacle Avoidance Task.

3. Joystick-Driven Obstacle Avoidance Task. Patients proceed towards a target and avoid oncoming obstacles.

See above Aravind & Lamontagne (2014). -

1. Obstacle Detection Task. Contralesional and head-on obstacles detected at closer proximities than ipsilesional obstacles.

2. Locomotor Obstacle Avoidance Task. Collisions only when approaching contralesional and head-on obstacles.

3. Joystick-Driven Obstacle Avoidance Task. For contralesional obstacles approaching, patients’ avoidance strategies start at shorter distances from the obstacle; minimum distances from contralesional obstacles are smaller than for ipsilesional obstacles.

Grattan & Woodbury (2017) RBD patients: 12 (6/6); 67.7 ± 8.8. RBD patients: 12 CVA; 617.5 d ± 852.7. BIT3’s Star Cancellation, Line Bisection subtests and all the BIT3’s behavioural subtest (BIT-bs); CBS7; Naturalistic Action Test (NAT)27. Buxbaum et al.’s (2012) VRALT’s enhanced array (distractor objects, auditory distractors, and moving distractors) in examiner condition (the examiner navigates the pathway using the joystick), because some participants are unable to operate the joystick. VRLAT run on a PC laptop using a Logitech Extreme 3D Pro joystick. - Neglect detected in six out of 12 RBD patients by both the conventional and the behavioural subtests of the BIT, in all 12 patients by the VRALT. VRALT scores correlate with CBS and NAT scores.
Aravind & Lamontagne (2018)

RBD (N+): 13; 59.8 ± 7.7.

RBD (N-): 13; 60.8 ± 6.5.

HC: 5 (3/5); age range: 50–74.

RBD (N+): 10.5 ± 4.6 m.

RBD (N-):11.8 ± 5.1 m.

Apples test1; Bells cancellation test5; letter cancellation test18; line bisection test22; MVPT26.

1. Locomotor Obstacle Avoidance Task.

2. Perceptual Task. See above Aravind & Lamontagne (2014).

See above Aravind & Lamontagne (2014). -

Collisions with left and head-on obstacles and longer latencies for detecting left vs. right obstacles by left RBD N + patients. More walking time spent by N + patients with head rotated rightwards, towards the ipsilesional side, as compared with RBD N- patients.

In RBD left N + patients, rightward deviations while walking in all obstacle conditions. N + patients fail to perform early or large enough adjustments to their heading and walking speed in response to contralesional obstacles, with increased risk of collisions.

Ogourtsova et al. (2018a, b)

RBD (N+): 12 (10/2); 60 ± 8.8.

RBD (N-): 15 (13/2); 58.5 ± 13.2.

HC: 9 (4/5); 56.3 ± 11.2.

RBD (N+): 12 CVA, 1,7 ± 1,1 y.

RBD (N-): 15 CVA, 2.0 ± 2.1 y.

Apples test1; Line Bisection22; Star Cancellation36.

Ecological VR-based Evaluation of Neglect Symptoms (EVENS). Immersive and 3D navigation test. Virtual presentation of a symmetrical and richly textured room, displaying a grocery shopping aisle with three shelves located in front and 3 m away from participants.

Target and distractors conditions.

1. Simple (target stand-alone in five possible locations).

2. Complex (target intermingled with distractors.

Tasks.

1. Detection (participants press the joystick button as soon as they detect the target).

2. Navigation (participants navigate using a joystick, controlling speed to reach the target with the most direct possible trajectory. Maximal Medio-Lateral Deviation (mMLD) from the best navigation trajectory computed.

VE created in the Unity® Game Engine. Viewing media: HMD, blocking all peripheral vision, with only VE visible to participants. Responses provided with the right hand, using a stationary joystick. Focus groups and questionnaire, directly involving expert clinicians, shows their interest and availability to use an immersive 3D tool. Detection task. In RBD N + patients, compared with RBD N- patients and HC, higher detection times for left-sided and middle targets.
Ogourtsova et al. (2018a, b)

RBD (N+): 15 (12/3); 60.2 ± 8.8.

RBD (N-): 15 (13/2); 58.5 ± 13.2.

HC: 15 (7/8); 61.0 ± 11.8

RBD (N+): 15 CVA, 1.6 ± 1.0 y.

RBD (N-): 15 CVA, 2.0 ± 2.1 y.

Apples Test 1; Line Bisection test (LBT)22; Star Cancellation test (SCT)36. LBT and SCT performed in the near (40 cm) and far (320 cm) space through a screen; participants’ responses provided by a laser pointer. VR-based Navigation and Detection tasks performed with participants seated and immersed in a 3D VE, representing a symmetrical and richly textured room, including a visual display of walls and ceiling. Target: a red ball, presented 7 m away from the starting navigation position in five possible locations, in different conditions: (1) remains visible during movement, (2) disappears after an initial auditory cue is presented, (3) moves rightwards or leftwards. VE created in Softimage XSI®. During trials, scene controlled by real-time CAREN-3™ software. Viewing media: a helmet-mounted display. Responses provided by joystick. -

In all VR-based Navigation conditions, more errors, overshooting left-sided targets, by RBD left N + patients vs. HC and RBD N- patients.

In the Detection task, longer detection times for all target location in RBD N + patients, vs. HC and RBD N- patients. In RBD N + patients, longer detection times for the L target compared to the

R target, as well as for

L and R targets compared to the middle ones.

Knobel et al. (2020)

RBD: 15 (7/8);

67.1 ± 10.05.

Of them, N+: 10; N-: 5.

HC: 35 (28/7); 69.0 ± 7.47.

RBD: 15 CVA;

76.5 ± 36.0 d.

Sensitive Neglect Test (SNT)30. Virtual Cancellation Task (VCT). A blue background and 120 objects (20 white-coloured spheres and 100 cubes), which changed their colour to red when touched with the controller. Participants find targets and reach and touch them using the hand-held controller and avoid distracters. VR setup: an HMD and a hand-held controller, connected to a gaming laptop.

Cybersickness rated very low

(SSQ)35. Very high ratings of acceptance and usability (SUS)38.

RBD N + patients perform worse than RBD N- patients and HC in both the VCT (higher total time and lower percent detected targets) and in the SNT. In RBD N + patients, CoC, displaced more leftward when measured by paper-and-pencil than by VCT. VCT detects neglect as reliably as paper-and-pencil cancellation tasks, then when measured by the VR cancellation task, the VR cancellation task could detect N as reliably as the used paper-pencil cancellation task.
Kim et al. (2021)

RBD (N+): 19 (14/5) 54.32 ± 7.40.

RBD (N-): 22 (17/4) 49.23 ± 9.99.

HC (10/12): 45.41 ± 17.82.

RBD (N+): 19 CVA, 11.79 ± 12.37 m.

RBD (N-): 22 CVA, 10.57 ± 10.96 m.

Line Bisection test (LBT)22; Star Cancellation test (SCT)36; CBS7.

FORP Task.

VR-based technique to assess Field of Perception (FOP) and Field of Regard (FOR) in N + patients. For FOP and FOR tasks, participants instructed to click the left or the right button of computer mouse, to a blue (left-sided click response) or red (right-sided response) target, presented in VE. FOP Task: targets detected in the absence of head movement. FOR Task. head movements for target detection allowed.

FOPR Tasks: stereo HMD system and 3D development platform controlled by desktop workstation running Windows 7, equipped with high-end graphics card. Screen resolution: 1280 × 800. Built-in 3-deg-of- freedom sensor to track head movements. - Differences detected by FOP and FOR Tasks in percent successful responses and response temporal indexes. RBD N + patients perform worse than both RBD N- patients and HC. Good agreement between FORP Tasks and paper-and-pencil Line Bisection and Star Cancellation tests and CBS.
OTHERS
Report N: Sex (M/F); Age (M ± SD)

Aetiology;

duration of disease

Conventional N Assessment General Description Technical aspects User experience Results
Sotokawa et al. (2015) RBD (N+): 3 (2/1). RBD (N+): 3 CVA. Japanese version of BIT4.

Driving-Simulator Lane Tracking task.

To keep the lane continuously while driving on a road with random curves. The ratio of straying from the course measured and calculated as error ratio. Duration of Driving-Simulator test: about 10 min.

Mitsubishi Inc.’s® Precision Driving Simulator (DS-20). -

Ratio of steering from the route biased leftwards. Interpretation: recovery from left neglect, with leftward overcompensation, possibly supported by pseudoneglect (i.e., the physiological minor leftward bias in line bisection, shown by healthy participants) (Jewell & McCourt, 2000).

No neglect at the BIT.

Peru et al. (2017)

RBD (N+): 6(2/4).

LBD (N+): 1 F.

(patients’ age range: 62–81).

HC: 7.

RBD (N+): 6 CVA.

LBD (N+): 1 CVA.

Bells Cancellation test5; BIT3’s Copying, Drawing and Line Bisection tests; Line Cancellation test24; Sentence Reading test31. A uniform distribution of stain images, differing in number (24 or 48) and features (coffee or water stains), and dust grains displayed on the top of an interactive table. Participants detect and cancel out stains using a round sponge. Stimuli reproduce stimulus size and spatial arrangement of papery Albert’s (1973) test. Each stain cancellation carried out under two conditions: (1) similar to papery Albert’s (1973) test (the spot, when “touched” by the sponge is marked by a red line); (2) more “ecologic” (the spot disappears, as the sponge touches it. Interactive table developed by University of Florence’s Media Integration and Communication Centre (MICC). Table device equipped with a 52 multi touch-screen monitor similar in shape to a dinner table with a wood surface-like. -

Flawless performance of 4 out of 7 RBD patients on the papery version of Albert’s (1973) test. Left-sided omission errors in the interactive table version by all seven patients. Patients’ performance overall better in the erasing than in the marking condition.

HC: errorless performance.

Montedoro et al. (2019)

Patients: 35 (24/11) (RBD: 23; LBD: 12); of them, 25 diagnosed with N+.

(patients’ mean age: 62.63 ± 11.89).

HC: 56, paired by mean age.

All the patients: 35 CVA, 11.17 ± 12.30 m. Apples test1; Bells Cancellation test5; Computerized Neglect subtest of the Test Battery for Attentional Performance (TAP)39.

MonAmour Task.

Target search task. Participants find a grey colored musical conductor (target), randomly intermingled among 119 black colored musical instruments of four different types (distractors). Targets complete or incomplete (in their left- or right-hand side), measuring then egocentric vs. allocentric neglect. Participants, when a complete target is detected, make a push response using the end-effector of the REAplan®robot towards the conductor’s baton; when an incomplete target is detected, they make a pull response, towards the conductor’s musical stand.

The REAplan®robot used for the development of the MonAmour test. REAplan®: a planar end-effector robot with position and force sensors allowing for precise action measures in the horizontal plane (dimensions 93 × 52-cm). -

For egocentric neglect, patients’ performance in computer-based task highly related to performance in paper-and-pencil tests, for left-right asymmetries in omissions and response latencies. Excellent sensitivity of MonAmour test, compared to both VRLAT (Buxbaum et al., 2006; Dawson et al., 2008) and VR-DiSTRO (Fordell et al., 2011).

For allocentric neglect, strong correlation between MonAmour test and the Apples test; better sensitivity and specificity of MonAmour test than the Apples test.

Cerrato et al. (2020)

RBD: 9 (6/3).

HC: 115 (54/61); 27.2 ± 8.6.

RBD: 9 CVA, 48.55 ± 55.33 m. Bells Cancellation, Copy of a Landscape Drawing, Overlapping Figure and Line Bisection tasks from the GEREN neglect battery15. E-TAN technology Platform. Enhanced version of Baking Tray Task2 (E-BTT). Tangible physical interfaces: 4-cm in diameter, to be placed on the platform. The polygon area defined by the spatial collocation of the disks on the platform indexes surface exploration. Timing and temporal sequence of collocations recorded. 30 fps camera used to detect disks’ position over the platform. Marker applied above each disk, recognized by ArUco Markers system. Some patients spontaneously describe the task as agreeable and quick to perform, and as a welcome change after so many boring paper-and-pencil tests. Great sensitivity of E-BTT to mild patial deficits, due to novel recorded measures, based on the use of the convex hull described by objects on the board. This measure provides an estimate of the portion of space processed by participants and may effectively discriminate N + vs. N- patients. No reported comparisons between performances in paper-and-pencil test vs. E-TAN.
Spreij et al. (2020)

RBD (N+): 33 (22/11); 58.83 ± 9.18.

LBD (N+): 7 (6/1); 54.75 ± 11.48.

RBD (recovered-N): 7 (4/3); 54.47 ± 14.69.

RBD and LBD (N-): 53 (40/13); 58.86 ± 12.14.

HC: 21 (11/10); 58.77 ± 9.86.

RBD (N+): CVA, 60.36 ± 31.83 d.

LBD (N+): CVA, 43.0 ± 26.54 d.

RBD (recovered-N): 7 CVA, 50.0 ± 37.76 d.

RBD and LBD (N-): 53 CVA, 41.89 ± 39.77 d.

Shape Cancellation task (SC)33: 54 targets shapes presented among 75 distractors on a computer screen monitor. Participants instructed to find all targets and to click on them: a circle appears on the screen around the location of the mouse click. CBS7.

Simulated driving task (van Kessel et al. 2010, 2013).

A driving scene projected on a large screen. No car interior shown; a steering wheel fixed on a table in front of participants, instructed to use it to maintain the starting position at the centre of the right lane. Simulated driving speed approximately 50 km/h at a set pace. Position in the lane, manipulated by simulated “side wind” from either side, needs to be adjusted continuously by participants. The projection of the driving scene vibrates as a warning signal, when participants drive off the road into the left or right verge. Task duration: two min.

Driving scene projected on a large screen (2.13  x 3.18-m), with participants seated in front of the screen, placed about 90-cm from their eyes. - Left N + RBD patients (assessment performed by the SC and the CBS) deviate leftward, contralesionally, towards the neglected side of the street, more than RBD N- patients. Interpretation: result possibly brought about by asymmetries in the layout of the simulated driving task.

Participants: Healthy Controls (HC); LBD/RBD (Left/Right-Brain-Damaged patients); MCI (patients with Mild Cognitive Impairment); (N+/-) Patients with/without spatial neglect. VE/VR (Virtual Environment/Reality).

Etiology and duration of disease: Cerebrovascular attack (CVA); brain tumor (BT); traumatic brain injury (TBI). d/w/m: days/weeks/months, years.

Conventional Tasks and Batteries: 1Apples test (Bickerton et al., 2011); 2Baking Tray Task (Tham, 1996); 3Behavioral Inattention Test – BIT (Wilson et al., 1987); 4Behavioral Inattention Test – BIT (Isiai, 1999); 5Bells cancellation test (Gauthier et al., 1989); 6Bells cancellation test (Isiai, 1999); 7Catherine Bergego Scale – CBS (Azouvi et al., 2003); 8Center of cancellation – CoC (Rorden & Karnath, 2010); 9Computer-Based Assessment of Visual Function – CAV (Niedeggen & Jörgens, 2005); 10Delis-Kaplan executive function system – DKEFS (Delis et al., 2012); 11Double Simultaneous Stimulation test – DSS (Bisiach & Faglioni, 1974; Làdavas, 1990); 12Drawing task (Ishiai et al., 1989); 13Figure copying task (Ogden, 1985); 14Fluff test (Cocchini et al., 2001); 15GEREN neglect battery (Azouvi et al., 2006); 16Geriatric Depression Scale – GDS (Yesavage & Sheikh, 1986); 17Hooper Visual Organization Test – HVOT (Hooper, 1983); 18Letter cancellation test (Diller & Weinberg, 1977); 19Letter cancellation test (Weintraub & Mesulam, 1988); 20Line bisection (Levine et al., 1986); 21Line bisection (Wilson et al., 1987); 22Line Bisection Test – LBT (Schenkenberg et al., 1980); 23Line Bisection Test – LBT (Lee et al., 2004); 24Line cancellation (Albert, 1973); 25Line cancellation (Levine et al., 1986); 26Motor Free Visual Perceptual Test – MVPT (Colarusso & Hammill, 1972); 27Naturalistic Action Test – NAT (Schwartz et al., 2002); 28Neglect-Test – NET (Fels & Geissner, 1997); 29Semi-Structured Scale For The Evaluation of Personal and Extrapersonal Neglect (Zoccolotti et al., 1992); 30Sensitive Neglect Test – SNT (Reinhart et al., 2016); 31Sentence reading (Barbieri & De Renzi, 1989); 32Sentence Reading (Zoccolotti et al., 1989); 33Shape Cancellation task – SC (Van der Stoep et al., 2013); 34Short Feedback Questionnaire (Rand et al., 2008); 35Simulator Sickness Questionnaire – SSQ (Kennedy et al., 1993); 36Star cancellation test (Wilson et al., 1987); 37Star cancellation test (Halligan et al., 1989); 38System Usability Scale - SUS (Brooke, 1996); 39test battery for attentional performance – TAP (Zimmermann & Fimm, 2002); 40Visual Extinction Test (Geeraerts et al., 2005); 41Wechsler Adult Intelligence Scale – WAIS-IV (Wechsler, 2012); 42Wechsler Memory Scale – WMS-IV (Wechsler, 2009); 43Western Aphasia Battery – WAB (Kertesz, 1982); 44Wundt-Jastrow illusion test (Massironi et al., 1988).

Analysis of the Quality of the Studies

The quality of the included studies was analyzed using the ad-hoc quality criteria of Corti et al. (2021). Four of the authors (ST, FA, GM and RE) independently evaluated the quality of the studies; no discrepancies in ratings among the evaluators were found. The evaluation criteria, based on parameters considered as relevant for performing experimental and clinical neuropsychological studies, considered the presence (1 = yes, 0 = no) of:

  1. Sample size calculation, based on power analysis, to avoid underpowered studies with unclear results, and the occurrence of publication bias. Recent literature points out the need to calculate the required sample size by using power analysis (Brysbaert, 2019; Kühberger et al., 2014). Specifically, for studies focused on tools for clinical neuropsychological assessment, power analysis can be performed by estimating the sample size needed to detect a significant difference in performance, with a specific power, with respect to a control group of participants not presenting impairments in the target construct, namely: healthy controls or patients without the impairment of interest (spatial neglect in the present review, see criterion below).

  2. Control group of healthy participants, of patients not presenting an impairment in the target construct, or of both. The comparison of the patients’ performances with those of a control group of participants, matched for age, sex, and educational level, is a standard neuropsychological practice, to evaluate the performance of a patients’ sample with a specific diagnosis (in this review, spatial neglect), with respect to participants who do not show that specific impairment, namely: healthy control participants, patients without spatial neglect or both (Kan et al., 2019; Vallar, 2000). Statistical corrections, such as using co-variates in the analysis of variance (Capitani & Laiacona, 1997), can be performed to control for the possible effects of concomitant variables (age, sex, educational level, duration of disease in stroke patients, severity of the overall neurological impairment, etc.) on the patients’ performances in the tests assessing the target construct of interest (in this review, neglect).

  3. Sensitivity or specificity reporting. Sensitivity is the ability of a test to correctly identify patients with a disease; specificity is the ability of a test to correctly identify people without the disease (Swift et al., 2020). Sensitivity and specificity are two psychometric properties that are used to assess target construct validity (i.e., the extent to which an instrument provides a measure of a theoretical construct).

  4. Measures of construct validity (the degree to which a test or instrument can measure a concept, trait, or other theoretical entity): convergent and divergent or discriminant validity reporting (Strauss & Smith, 2009). Convergent or congruent validity is the extent to which responses to a test or to an instrument exhibit a strong relationship with responses to similar tests or instruments, namely the extent to which a measure correlates with other measures of the same or of a similar construct. Divergent or discriminant validity is the degree to which a test or measure diverges from (i.e., does not correlate with) another measure, whose underlying construct is conceptually unrelated to it.

  5. Ecological validity reporting. Ecological validity indicates if the findings of a study can be generalized to realistic situations in everyday life. It is a subtype of external validity, that can be defined as the extent to which the measures of outcome correlate with, or predict, performance in ADLs (see also Holleman et al., 2020).

  6. User experience reporting. This criterion was added to those proposed by Corti et al. (2021), to evaluate the presence of negative (i.e., heavy burden or adverse reactions, as headache, fatigue or boredom) or positive (i.e., feelings of fun and safety) reporting about the experience of using these digital tools, as compared to the traditional paper-and-pencil ones. This criterion was considered as met if user experience reporting was obtained from healthy participants, brain-damaged patients, clinicians, or from all of them.

The analysis of the quality of each study was then based on these six ad-hoc established criteria. With respect to sample size calculation based on power analysis, only two studies (Aravind & Lamontagne, 2014; Quinn et al., 2018) applied this procedure (4.76%). Twenty-six studies included a control group of healthy participants (61.90%), 19 had a control group of patients without spatial neglect (45.23%), and 12 included both (28.57%). As for psychometric properties, 23 studies reported sensitivity or specificity measures (45.23%), 20 convergent or discriminant validity (46.61%) and nine ecological validity (14.28%). Furthermore, among the studies reporting specificity or sensitivity, only two (Buxbaum et al., 2012; Chung et al., 2016) also reported the cut-off for impairment. This means that only for two out of 42 studies data to classify the patients’ performance as impaired or non-impaired (4.76%) were available. Finally, with respect to user experience reporting, only nine studies showed these data (21.42%). Table 2 shows the analysis of the quality of each study, based on the six criteria reported above.

Table 2.

Evaluation of the quality of the studies

COMPUTER
Report Sample size calculation based on power analysis Presence of healthy control group Presence of N- control group Sensitivity/Specificity reporting Convergent Validity/Discriminant Validity reporting Ecological Validity reporting User experience report
Chiba et al. (2010) 0 0 0 0 0 0 0
Rabuffetti et al. (2012) 0 1 1 1 0 0 0
Bonato et al. (2013) 0 0 0 0 0 0 0
Ulm et al. (2013) 0 1 0 1 1 0 1
Jee et al. (2015) 0 1 0 1 0 0 0
Blini et al. (2016) 0 1 0 0 0 0 0
Mizuno et al. (2016) 0 1 1 0 0 0 0
Machner et al. (2018) 0 1 1 1 1 1 0
Rossa et al. (2019) 0 1 0 0 0 0 0
Kocanaogullari et al. (2020) 0 0 1 1 0 0 0
Villarreal et al. (2020) 0 1 0 1 0 1 0
Villarreal et al. (2021) 0 1 0 1 0 0 0
Villarreal et al. (2022) 0 1 0 1 0 0 0
Total studies: N = 13 0/13 (0%) 10/13 (76.92%) 4/13 (30.76%) 8/13 (61.53%) 2/13 (15.38%) 2/13 (15.38%) 1/13 (7.69%)
GRAPHICS TABLET AND TOUCH-SCREEN TABLET
Report Sample size calculation based on power analysis Presence of healthy control group Presence of N- control group Sensitivity/Specificity reporting Convergent Validity/Discriminant Validity reporting Ecological Validity reporting User experience report
Liang et al. (2010) 0 0 1 1 1 0 0
Smit et al. (2013) 0 0 1 0 0 0 0
Pallavicini et al. (2015) 0 0 1 1 0 1 1
Vaes et al. (2015) 0 1 0 0 0 0 0
Chung et al. (2016) 0 1 1 1 1 0 0
Willer et al. (2016) 0 1 0 0 1 0 0
Quinn et al. (2018) 1 0 0 1 0 0 1
Pierce et al. (2021) 0 1 1 1 1 0 0
Total studies: N = 8 1/8 (12.5%) 4/8 (50%) 5/8 (62.5%) 5/8 (62.5%) 4/8 (50%) 1/8 (12.5%) 2/8 (25%)
VIRTUAL REALITY
Report Sample size calculation based on power analysis Presence of healthy control group Presence of N- control group Sensitivity/Specificity reporting Convergent Validity/Discriminant Validity reporting Ecological Validity reporting User experience report
Kim et al. (2010) 0 0 1 0 1 0 0
Tanaka (2010) 0 0 0 0 1 0 0
Fordell et al. (2011) 0 0 1 1 1 0 1
Mesa-Gresa et al. (2011) 0 0 1 0 1 0 0
Peskine et al. (2011) 0 1 0 0 0 0 0
Sugarman et al. (2011) 0 0 0 0 0 0 1
Buxbaum et al. (2012) 0 1 0 1 1 1 1
Dvorkin et al. (2012) 0 1 1 1 0 0 0
Aravind & Lamontagne (2014) 1 0 0 0 1 0 0
Aravind et al. (2015) 0 0 0 0 1 0 0
Grattan & Woodbury (2017) 0 0 0 1 0 0 0
Aravind & Lamontagne (2018) 0 1 1 0 1 0 0
Ogourtsova et al. (2018a, b) 0 1 1 1 1 0 1
Ogourtsova et al. (2018a, b) 0 1 1 0 1 0 0
Knobel et al. (2020) 0 1 1 1 1 0 1
Kim et al. (2021) 0 1 1 1 1 1 0
Total studies: N = 16 1/16 (6.25%) 8/16 (50%) 9/16 (56.25%) 7/16 (43.75%) 12/16 (75%) 2/16 (12.5%) 5/16 (31.25%)
OTHERS
Report Sample size calculation based on power analysis Presence of healthy control group Presence of N- control group Sensitivity/Specificity reporting Convergent Validity/Discriminant Validity reporting Ecological Validity reporting User experience report
Sotokawa et al. (2015) 0 0 0 0 0 0 0
Peru et al. (2017) 0 1 0 1 0 0 0
Montedoro et al. (2019) 0 1 0 1 1 0 0
Cerrato et al. (2020) 0 1 0 0 0 0 1
Spreij et al. (2020) 0 1 1 1 1 1 0
Total studies: N = 5 0/5 (0%) 4/5 (80%) 1/5 (20%) 3/5 (60%) 2/5 (40%) 1/5 (20%) 1/5 (20%)
Overall: N = 42 2/42 (4.76%) 26/42 (61.90%) 19/42 (45.23%) 23/42 (54.76%) 20/42 (47.61%) 6/42 (14.28%) 9/42 (21.42%)

Discussion

At present, the main manifestations of spatial neglect are mainly detected through clinical observation and neuropsychological paper-and-pencil tests, that allow to explore its features (e.g., egocentric vs. allocentric, personal vs. extra-personal) in a standardized way, and to quantify its severity (Heilman et al., 2003; Lezak et al., 1995; Vallar & Bolognini, 2014). However, some pieces of evidence point out that, the longer the duration of the disease after stroke onset, the better the patients’ performances in neuropsychological tests, although difficulties are still found with ADLs (Bonato et al., 2010, 2012; Della Sala et al., 2018). This improvement of spatial neglect in stroke patients is due to both spontaneous recovery of neurological and neuropsychological deficits (Corbetta et al., 2005; Murphy & Corbett, 2009; Stefaniak et al., 2020; Stinear, 2017), and the effects of rehabilitation (Azouvi et al., 2017; Bowen et al., 2013; Kerkhoff & Schenk, 2012). Emerging literature suggests that IT-based devices could be an important tool for neuropsychological assessment. These technologies allow to record much more information with respect to paper-and-pencil tests, and to replicate and alter the surrounding world, permitting to evaluate the patients’ behavior in contexts resembling real life conditions (Ogourtsova et al., 2017; Pedroli et al., 2015; Tsirlin et al., 2009). Despite the potential of these technologies, several issues need to be taken into account when developing, disseminating and implementing CNADs (Bauer et al., 2012; see Kourtesis & MacPherson, 2021 for VR-related issues).

This scoping review aimed at presenting an overview of the studies in which IT-based devices were used as tools to assess manifestations of spatial neglect. The inclusion criteria were met by 42 articles, which were categorized into five groups, depending on their technological approaches: computer (13), graphics tablet or tablet (8), virtual reality-based assessment (16), and other (5).

Computer (pc/laptop)

Computer tools have proved to be effective for the evaluation of manifestations of spatial neglect. These systems appear to be more sensitive in recording difficulties along the left-right spatial dimension, with respect to the gold standard and time-honoured paper-and-pencil tests. Many advantages of this kind of technology are acknowledged by the AACN and the NAN, as well as by this review. Relevant issues include the following: (i) testing large numbers of participants quickly (e.g., parallel administration); (ii) tests available at any time; (iii) enhanced accuracy and precision (e.g., reaction times, or exploration time of the two sides of space, Machner et al., 2018; Mizuno et al., 2016; Rabuffetti et al., 2012; Rossa et al., 2019; Villarreal et al., 2020, 2021); (iv) shorter administration times and reduced costs for test administration and scoring; (v) when appropriate, test versions for different languages; (vi) automatically exporting recorded data to databases (Chiba et al., 2010; Jee et al., 2015, for a digital version of the line bisection task); (vii) increased accessibility, for instance from remote; (viii) integration of algorithms for making decisions on issues such as the identification of an impairment or of a statistically reliable change in performance. These technologies can be easily integrated with other systems, thus providing a wider range of measures (e.g., Electroencephalography, Kocanaogullari et al., 2020). Finally, computerized tests adequately suit the purpose of being administered via telematics from remote: the currently ongoing SARS-CoV-2 world pandemic has emphasized this need.

On the other hand, the present qualitative evaluation of the studies points out several critical issues that must be addressed. With respect to sample size calculation, none of the studies applied power analyses. Ten studies included a control group of healthy participants, and four studies a control group of patients without spatial neglect. Only three out of 13 studies included both. This means that the ability of about 75% of the studies to distinguish different levels of performance among different groups of participants (healthy, affected or not affected by the deficit under investigation, as indicated by other tasks) was not completely assessed, making uncertain the possible clinical validity of most of the proposed IT-based tasks for distinguishing a defective from a preserved performance. As for psychometric properties, eight studies reported sensitivity or specificity measures; two studies report convergent or discriminant validity and only one ecological validity (only Machner et al., (2018) reported both). None of the studies reporting specificity or sensitivity, provided cut-offs for detecting performance impairments. Finally, with respect to user experience reporting, only Ulm et al. (2013) showed these data. The parameters of ecological validity and user experience reporting are most relevant, considering that ADL scales are most sensitive to detect spatial neglect (Azouvi, 2017) and the used IT-based devices are novel tools for both clinicians and patients.

Graphic Tablet and Touch-Screen Tablet

The same list of pros discussed in the previous section readily applies to graphics and touch-screen tablets. The present qualitative evaluation raises several critical issues also for the studies included in this group. In the first place, sample size calculation based on power analysis was applied by only one study (Quinn et al., 2018). A control group of healthy participants was included in four studies and five studies included a control group of patients without spatial neglect; only two out of eight studies had both. This means, as for the computer group, that in about 75% of the studies, the possible clinical validity of most of the proposed IT-based tasks for distinguishing a defective from a preserved performance is uncertain. As for psychometric properties, sensitivity, or specificity measures and convergent or discriminant validity were more frequently reported than in the computer group: five studies reported sensitivity or specificity measures (62.5%) however, among them, only Chung et al. (2016) reported cut-offs for detecting impaired performances. Four studies (50%) reported convergent or discriminant validity. Ecological validity was underreported also in this group: only the article by Pallavicini et al. (2015) provided these measures. Finally, only two out of eight studies mentioned user experience reporting (Pallavicini et al., 2015; Quinn et al., 2018), showing high ratings of acceptance and usability of the NeglectApp and of the StrokeVision app.

Virtual Reality

Among the 16 studies included in this group, non-immersive paradigms were used only in two studies (Buxbaum et al., 2012; Grattan & Woodbury, 2017). The Virtual Reality Lateralized Attention Test (VRLAT) consists indeed of a 2-dimensional screen presentation of a virtual, nonbranching path, in which participants interact with the environment using a joystick. The VRLAT detects manifestation of spatial neglect better than the traditional paper-and-pencil tasks; moreover, scores obtained at the VRLAT correlate with those obtained in the CBS and in the Naturalistic Action Test (Schwartz et al., 2002). However, Buxbaum et al. (2012) report that six out of 70 right brain-damaged patients (about 9%) did not complete the experimental procedure, due to difficulties in operating with the joystick; no other measures of user experience are provided. All the remaining 14 studies included in the VR group employed instead immersive paradigms: in particular, 10 studies integrated computers with head-mounted display, two used shutter glasses (Dvorkin et al., 2012; Fordell et al., 2011), and two body-tracking sensors (Mesa-Gresa et al., 2011; Sugarman et al., 2011). Regarding the methodological aspects, only Aravind and Lamongagne (2014) applied sample size calculation based on power analysis. Eight studies included a control group of healthy participants and eight a control group of patients without spatial neglect; six studies included both groups. As for psychometric properties, seven studies (43.75%) reported sensitivity or specificity measures, but only Buxbaum et al. (2012), also cut-offs for detecting performance impairments. Twelve studies reported instead convergent or discriminant validity measures (75%).

VR paradigms allow to replicate and alter the surrounding world to evaluate the patients’ behavior in contexts simulating real life (Ogourtsova et al., 2017; Pedroli et al., 2015; Tsirlin et al., 2009). However, only Buxbaum et al. (2012), who employed a non-immersive paradigm, and Kim et al. (2021), who used instead an immersive paradigm, provided ecological validity measures (12.5%). The paucity of measures regarding ecological validity is worth noticing, considering that one possible advantage of VR paradigms is to create scenarios that resemble real life situations, thus giving leeway to the evaluation of the patients’ behaviour in these contexts. In particular, the fact that only one out of 14 studies adopting immersive paradigms reported these measures is even more notable, considering that immersive VR technology allows to collect data through the employment of dynamic stimuli and interactions with a high degree of control within an ecologically valid environment (Kourtesis et al., 2020; Rizzo et al., 2004). Finally, with respect to user experience ratings, data, although encouraging, are few: as for non-immersive VR-paradigms, about 9% of the clinical sample did not complete Buxbaum et al.’s (2012) test, due to difficulties in operating with the joystick. As for the immersive VR-paradigms, Fordell et al. (2011) employed a three-item questionnaire: results show that, during the VR-DiSTRO battery administration, 41% of the clinical sample reported to feel focused, 22% to feel pleased during the administration of the task, and 25% to feel alert. Sugarman et al. (2011) made use of the Short Feedback Questionnaire (Rand et al., 2008): the only patient tested with the VR react task of the SeeMe system reported feelings of enjoy, control and success in the VR environment. In the study by Knobel et al. (2020), the administration of the System Usability Scale (Brooke, 1996) to the clinical sample revealed high ratings of acceptance and usability of their VR cancellation task. Conversely, Ogourtsova et al. (2018a, b) conducted focus groups and administered questionnaires directly involving the clinician: experts showed interest and availability to use VR paradigms in their clinical practice. Finally, only two studies investigated VRISE: no participant reported nausea or other adverse effects during or after the administration of the VR-DiSTRO battery of Fordell et al. (2011). Furthermore, cybersickness connected to the VR cancellation task of Knobel et al.‘s (2020) was rated as “very low” at the Simulator Sickness Questionnaire (Kennedy et al., 1993).

Other

Studies included in this section employ an interactive table (Peru et al., 2017), a planar end-effector robot (REAplan®robot) (Montedoro et al., 2019), a tangible user interface system consisting in an integrated system of concrete objects that participants can manipulate (E-TAN platform) (Cerrato et al., 2020) and a driving simulator (Sotokawa et al., 2015; Spreij et al., 2020). None of the included study applied sample size calculation procedures. Four out of five studies included a control group of healthy participants, but only one (Spreij et al., 2020) also a control group of patients without neglect. Sensitivity or specificity measures were reported in three studies, convergent or discriminant validity measures in two. Only Spreij et al. (2020) reported measures of ecological validity. Finally, with respect to user experience, only Cerrato et al. (2020) provided the reports of some patients, who spontaneously described the task as agreeable and quick to perform, and as a welcome change after so many boring paper-and-pencil tests.

For sure, the simulated driving-based assessment offers an evaluation based on tasks like the situations people face in real life while driving; however, the two studies included in this section share the same limitations previously mentioned for VR paradigms. Furthermore, a requirement of driving simulators and VR devices is that both the experimenter and the participant have to be trained to be able to use them; this can be difficult for patients with cognitive impairment, such as memory deficits (Baddeley et al., 2003; Ferbinteanu, 2019). The same problem concerns all the innovative devices included in this section, that are realized with proprietary hardware and software components, since a specific training for each technological proposal is required.

Taken together, data emerging from the present qualitative synthesis of the studies included in this scoping review, although encouraging, are still few and leave than room for several critical issues. In the first place, from a methodological point of view, the recent literature points out the need to calculate the required sample size by using power analysis (Brysbaert, 2019; Kühberger et al., 2014). Accordingly, underpowered studies which may provide unclear results, and then the occurrence of publication biases, are to be avoided. However, despite the above mentioned methodolological critical issues, only two studies included in our review (about the 5%) used sample size calculation procedures. Furthermore, only 12 studies (less than 30%) included both a control group of healthy participants and patients without neglect: the comparison of the patients’ performances with those of a control group of participants, is indeed standard neuropsychological practice, to evaluate the performance of a patients’ sample with a specific diagnosis (in this review, spatial neglect), with respect to participants who do not show that specific impairment, namely: healthy control participants, patients without spatial neglect or both (Kan et al., 2019; Vallar, 2000). As for psychometric features, sensitivity, sensibility measures and convergent or divergent validity were reported in about 50% of the reviewed studies. Moreover, considering that in the subacute, and particularly in the chronic phase post-stroke, patients may still show difficulties in ADLs despite doing well in neuropsychological paper-and-pencil tests (Bonato et al., 2010, 2012; Della Sala et al., 2018), it is even more notable that ecological validity was addressed only in about 13% of the studies. Ecological validity is a subtype of external validity, that can be defined as the extent to which the measures of outcome correlate with, or predict, performance in ADLs (see also Holleman et al. 2020). Finally, despite user experience represents a further critical issue that should be addressed to benefit from the advantages of CNADs (Bauer et al., 2012; see Kourtesis and MacPherson, 2021 for VR-related issues), only about the 20% of the included studies reported user experience measures.

Conclusion

The current findings make it difficult to establish golden standard tests or assessment procedures other than those (i.e., paper-and-pencil tests), that are extensively used in clinical settings for the evaluation of spatial neglect. Moreover, the process of developing IT-based tests in clinical neuropsychology should take into account the issues raised for the development, implementation and dissemination of CNADs and VR paradigms (Bauer et al., 2012; see Kourtesis & MacPherson, 2021 for VR-related issues). This is a laborious procedure, requiring the cooperation of a multidisciplinary team, including clinicians, researchers, informatics, interaction designers and engineers. While embracing technology is necessary, it is also essential to be aware of every aspect that leads to an efficient and effective assessment (see, for instance, Kourtesis & MacPherson, 2021). Nevertheless, it is our view that, through technical and user experience improvements, studies in brain-damaged patients and control participants, comparing paper-and-pencil and IT-based devices, and normative studies providing the cut-off scores, needed for the diagnostic process (e.g., Mancuso et al., 2015; Vallar et al., 1994), it will be possible to gather increasing evidence about the efficacy of at least some of the tests considered in this review. This, in turn, will provide additional support for changes and improvements in clinical practice, for diagnostic and follow up purposes. This review and the discussion of the available IT-based studies provides a database of information for the future developments of this line of research.

Credit Authorship Contribution Statement

Stefano Terruzzi: Methodology, Investigation, Visualization, Data curation, Writing - Original draft, Supervision. Federica Albini: Investigation, Visualization, Data curation, Writing - Original draft. Gemma Massetti: Investigation, Visualization, Data curation, Writing - Original draft. Roberta Etzi: Investigation, Visualization, Data curation, Writing - Review & Editing. Alberto Gallace: Conceptualization, Methodology, Writing - Review & Editing, Supervision, Funding acquisition. Giuseppe Vallar: Conceptualization, Methodology, Writing - Review & Editing, Supervision, Project administration, Funding acquisition.

Funding

This research received funding to Giuseppe Vallar from the Fondo di Ateneo per la Ricerca of the University of Milano Bicocca and from the Fondi per la Ricerca Corrente by the IRCCS Istituto Auxologico Italiano. This research received also funding to Alberto Gallace from the Fondo di Ateneo per la Ricerca of the University of Milano Bicocca.

Open access funding provided by Università degli Studi di Milano - Bicocca within the CRUI-CARE Agreement.

Data Availability

data sharing not applicable to this article as no datasets were generated or analyzed during the current study.

Declarations

Declaration of Interest

none.

Footnotes

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Contributor Information

Stefano Terruzzi, Email: s.terruzzi2@campus.unimib.it.

Giuseppe Vallar, Email: giuseppe.vallar@unimib.it.

References

  1. Albert ML. A simple test of visual neglect. Neurology. 1973;23(6):658–658. doi: 10.1212/WNL.23.6.658. [DOI] [PubMed] [Google Scholar]
  2. Antonucci G, Guariglia C, Judica A, Magnotti L, Paolucci S, Pizzamiglio L, Zoccolotti P. Effectiveness of neglect rehabilitation in a randomized group study. Journal of Clinical and Experimental Neuropsychology. 1995;17(3):383–389. doi: 10.1080/01688639508405131. [DOI] [PubMed] [Google Scholar]
  3. Aravind G, Darekar A, Fung J, Lamontagne A. Virtual reality-based Navigation Task to reveal obstacle avoidance performance in individuals with Visuospatial Neglect. IEEE Transactions on Neural Systems and Rehabilitation Engineering. 2015;23(2):179–188. doi: 10.1109/TNSRE.2014.2369812. [DOI] [PubMed] [Google Scholar]
  4. Aravind G, Lamontagne A. Perceptual and locomotor factors affect obstacle avoidance in persons with visuospatial neglect. Journal of NeuroEngineering and Rehabilitation. 2014;11(1):38. doi: 10.1186/1743-0003-11-38. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Aravind G, Lamontagne A. Effect of visuospatial neglect on spatial navigation and heading after stroke. Annals of Physical and Rehabilitation Medicine. 2018;61(4):197–206. doi: 10.1016/j.rehab.2017.05.002. [DOI] [PubMed] [Google Scholar]
  6. Arduino LS, Burani C, Vallar G. Lexical decision in Left Neglect Dyslexia: Word frequency and Nonword Orthographic Neighborhood Effects. Cortex; A Journal Devoted To The Study Of The Nervous System And Behavior. 2002;38(5):829–832. doi: 10.1016/S0010-9452(08)70050-9. [DOI] [Google Scholar]
  7. Arrigoni G, De Renzi E. Constructional apraxia and hemispheric locus of lesion. Cortex; A Journal Devoted To The Study Of The Nervous System And Behavior. 1964;1(2):170–197. doi: 10.1016/S0010-9452(64)80020-4. [DOI] [Google Scholar]
  8. Azouvi P. Functional consequences and awareness of unilateral neglect: study of an evaluation scale. Neuropsychological Rehabilitation. 1996;6(2):133–150. doi: 10.1080/713755501. [DOI] [Google Scholar]
  9. Azouvi P. Sensitivity of clinical and behavioural tests of spatial neglect after right hemisphere stroke. Journal of Neurology Neurosurgery & Psychiatry. 2002;73(2):160–166. doi: 10.1136/jnnp.73.2.160. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Azouvi P. The ecological assessment of unilateral neglect. Annals of Physical and Rehabilitation Medicine. 2017;60(3):186–190. doi: 10.1016/j.rehab.2015.12.005. [DOI] [PubMed] [Google Scholar]
  11. Azouvi P, Bartolomeo P, Beis JM, Perennou D, Pradat-Diehl P, Rousseaux M. A battery of tests for the quantitative assessment of unilateral neglect. Restorative neurology and neuroscience. 2006;24(4–6):273–285. [PubMed] [Google Scholar]
  12. Azouvi P, Jacquin-Courtois S, Luauté J. Rehabilitation of unilateral neglect: evidence-based medicine. Annals of physical and rehabilitation medicine. 2017;60(3):191–197. doi: 10.1016/j.rehab.2016.10.006. [DOI] [PubMed] [Google Scholar]
  13. Azouvi P, Olivier S, de Montety G, Samuel C, Louis-Dreyfus A, Tesio L. Behavioral assessment of unilateral neglect: study of the psychometric properties of the Catherine Bergego Scale. Archives of Physical Medicine and Rehabilitation. 2003;84(1):51–57. doi: 10.1053/apmr.2003.50062. [DOI] [PubMed] [Google Scholar]
  14. Baddeley, A. D., Kopelman, M. D., & Wilson, B. A. (2003). The handbook of memory disorders. John Wiley & Sons.
  15. Barbieri C, De Renzi E. Patterns of neglect dissociation. Behavioural Neurology. 1989;2(1):13–24. doi: 10.1155/1989/728487. [DOI] [PubMed] [Google Scholar]
  16. Barer DH. The influence of visual and tactile inattention on predictions for recovery from acute stroke. QJM: An International Journal of Medicine. 1990;74(1):21–32. [PubMed] [Google Scholar]
  17. Barrett AM, Houston KE. Update on the Clinical Approach to spatial neglect. Current Neurology and Neuroscience Reports. 2019;19(5):25. doi: 10.1007/s11910-019-0940-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Batchelor FA, Mackintosh SF, Said CM, Hill KD. Falls after stroke. International Journal of Stroke. 2012;7(6):482–490. doi: 10.1111/j.1747-4949.2012.00796.x. [DOI] [PubMed] [Google Scholar]
  19. Bauer RM, Iverson GL, Cernich AN, Binder LM, Ruff RM, Naugle RI. Computerized neuropsychological assessment devices: joint position paper of the American Academy of Clinical Neuropsychology and the National Academy of Neuropsychology. Archives of Clinical Neuropsychology. 2012;27(3):362–373. doi: 10.1093/arclin/acs027. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Bergego, C., Azouvi, P., Samuel, C., Marchal, F., Louis-Dreyfus, A., Jokic, C. (1995). Validation d’une échelle d’évaluation fonctionnelle de l’héminégligence dans la vie quotidienne: l’échelle CB. Annales de Réadaptation et de Médecine Physique, 38(4), 183–189. 10.1016/0168-6054(96)89317-2
  21. Bernspång B, Asplund K, Eriksson S, Fugl-Meyer AR. Motor and perceptual impairments in acute stroke patients: effects on self-care ability. Stroke. 1987;18(6):1081–1086. doi: 10.1161/01.STR.18.6.1081. [DOI] [PubMed] [Google Scholar]
  22. Bickerton WL, Samson D, Williamson J, Humphreys GW. Separating forms of neglect using the Apples Test: validation and functional prediction in chronic and acute stroke. Neuropsychology. 2011;25(5):567–580. doi: 10.1037/a0023501. [DOI] [PubMed] [Google Scholar]
  23. Bisiach E, Bulgarelli C, Sterzi R, Vallar G. Line bisection and cognitive plasticity of unilateral neglect of space. Brain and Cognition. 1983;2(1):32–38. doi: 10.1016/0278-2626(83)90027-1. [DOI] [PubMed] [Google Scholar]
  24. Bisiach E, Capitani E, Luzzatti C, Perani D. Brain and conscious representation of outside reality. Neuropsychologia. 1981;19(4):543–551. doi: 10.1016/0028-3932(81)90020-8. [DOI] [PubMed] [Google Scholar]
  25. Bisiach E, Cornacchia L, Sterzi R, Vallar G. Disorders of perceived auditory lateralization after lesionsof the right hemisphere. Brain. 1984;107(1):37–52. doi: 10.1093/brain/107.1.37. [DOI] [PubMed] [Google Scholar]
  26. Bisiach E, Faglioni P. Recognition of random shapes by patients with unilateral lesions as a function of complexity, association value and delay. Cortex; A Journal Devoted To The Study Of The Nervous System And Behavior. 1974;10(2):101–110. doi: 10.1016/S0010-9452(74)80001-8. [DOI] [PubMed] [Google Scholar]
  27. Bisiach E, Luzzatti C. Unilateral neglect of Representational Space. Cortex; A Journal Devoted To The Study Of The Nervous System And Behavior. 1978;14(1):129–133. doi: 10.1016/S0010-9452(78)80016-1. [DOI] [PubMed] [Google Scholar]
  28. Bisiach E, Luzzatti C, Perani D. Unilateral neglect, representational schema and consciousness. Brain: a journal of neurology. 1979;102(3):609–618. doi: 10.1093/brain/102.3.609. [DOI] [PubMed] [Google Scholar]
  29. Bisiach E, Perani D, Vallar G, Berti AM. Unilateral neglect: personal and extra-personal. Neuropsychologia. 1986;24(6):759–767. doi: 10.1016/0028-3932(86)90075-8. [DOI] [PubMed] [Google Scholar]
  30. Bisiach, E., & Vallar, G. (2000). Unilateral neglect in humans. In F. Boller, J. Grafman, & G. Rizzolatti (Eds.), Handbook of Neuropsychology (1 vol., pp. 459–502). Elsevier Science Publishers BV.
  31. Blini E, Romeo Z, Spironelli C, Pitteri M, Meneghello F, Bonato M, Zorzi M. Multi-tasking uncovers right spatial neglect and extinction in chronic left-hemisphere stroke patients. Neuropsychologia. 2016;92:147–157. doi: 10.1016/j.neuropsychologia.2016.02.028. [DOI] [PubMed] [Google Scholar]
  32. Bonato M, Deouell LY. Hemispatial Neglect: computer-based testing allows more sensitive quantification of Attentional Disorders and Recovery and might lead to Better evaluation of Rehabilitation. Frontiers in Human Neuroscience. 2013 doi: 10.3389/fnhum.2013.00162. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Bonato M, Priftis K, Marenzi R, Umiltà C, Zorzi M. Increased attentional demands impair contralesional space awareness following stroke. Neuropsychologia. 2010;48(13):3934–3940. doi: 10.1016/j.neuropsychologia.2010.08.022. [DOI] [PubMed] [Google Scholar]
  34. Bonato M, Priftis K, Marenzi R, Umiltà C, Zorzi M. Deficits of contralesional awareness: a case study on what paper-and-pencil tests neglect. Neuropsychology. 2012;26(1):20–36. doi: 10.1037/a0025306. [DOI] [PubMed] [Google Scholar]
  35. Bonato M, Priftis K, Umiltà C, Zorzi M. Computer-based attention-demanding testing unveils severe neglect in apparently intact patients. Behavioural Neurology. 2013;26(3):179–181. doi: 10.1155/2013/139812. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Bowen A, Hazelton C, Pollock A, Lincoln NB. Cognitive rehabilitation for spatial neglect following stroke. The Cochrane Database of Systematic Reviews. 2013 doi: 10.1002/14651858.CD003586.pub3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Bowen A, McKenna K, Tallis RC. Reasons for variability in the reported rate of occurrence of unilateral spatial neglect after stroke. Stroke. 1999;30(6):1196–1202. doi: 10.1161/01.STR.30.6.1196. [DOI] [PubMed] [Google Scholar]
  38. Brain WR. Visual disorientation with special reference to lesions of the right cerebral hemisphere. Brain. 1941;64(4):244–272. doi: 10.1093/brain/64.4.244. [DOI] [Google Scholar]
  39. Broca, P. (1861). Remarques sur le siège de la faculté du langage articulé, suivies d’une observation d’aphémie (perte de la parole) (6 vol., pp. 330–357). Bulletin et Memoires de la Societe anatomique de Paris.
  40. Brooke, J. (1996). Sus: a “quick and dirty’usability. Usability evaluation in industry, 189.
  41. Brysbaert M. How many participants do we have to include in properly powered experiments? A tutorial of power analysis with reference tables. Journal of Cognition. 2019;2(1):16. doi: 10.5334/joc.72. [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Bucur M, Papagno C. Are transcranial brain stimulation effects long-lasting in post-stroke aphasia? A comparative systematic review and meta-analysis on naming performance. Neuroscience & Biobehavioral Reviews. 2019;102:264–289. doi: 10.1016/j.neubiorev.2019.04.019. [DOI] [PubMed] [Google Scholar]
  43. Bucur M, Papagno C. An ALE meta-analytical review of the neural correlates of abstract and concrete words. Scientific Reports. 2021;11(1):15727. doi: 10.1038/s41598-021-94506-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Bucur M, Papagno C. Deep brain stimulation in Parkinson Disease: a Meta-analysis of the long-term neuropsychological outcomes. Neuropsychology Review. 2022 doi: 10.1007/s11065-022-09540-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Buxbaum LJ, Dawson AM, Linsley D. Reliability and validity of the virtual reality lateralized attention test in assessing hemispatial neglect in right-hemisphere stroke. Neuropsychology. 2012;26(4):430–441. doi: 10.1037/a0028674. [DOI] [PubMed] [Google Scholar]
  46. Buxbaum LJ, Ferraro MK, Veramonti T, Farne A, Whyte J, Ladavas E, et al. Hemispatial neglect: subtypes, neuroanatomy, and disability. Neurology. 2004;62(5):749–756. doi: 10.1212/01.WNL.0000113730.73031.F4. [DOI] [PubMed] [Google Scholar]
  47. Buxbaum LJ, Palermo MA, Mastrogiovanni D, Read MS, Rosenberg-Pitonyak E, Rizzo AA, Coslett HB. Assessment of spatial attention and neglect with a virtual wheelchair navigation task. Journal of Clinical and Experimental Neuropsychology. 2008;30(6):650–660. doi: 10.1080/13803390701625821. [DOI] [PubMed] [Google Scholar]
  48. Buxbaum, L. J., Palermo, M., Mastrogiovanni, D., Read, M. S., Rosenberg-Pitonyak, E., Rizzo, A. A., & Coslett, H. B. (2006). Assessment of Spatial Neglect with a Virtual Wheelchair Navigation Task. In 2006 International Workshop on Virtual Rehabilitation (pp. 94–99). Presented at the 2006 International Workshop on Virtual Rehabilitation, New York, NY: IEEE. 10.1109/IWVR.2006.1707534
  49. Capitani E, Laiacona M. Composite neuropsychological batteries and demographic correction: standardization based on equivalent scores, with a review of published data. Journal of Clinical and Experimental Neuropsychology. 1997;19(6):795–809. doi: 10.1080/01688639708403761. [DOI] [PubMed] [Google Scholar]
  50. Cerrato, A., Pacella, D., Palumbo, F., Beauvais, D., Ponticorvo, M., Miglino, O., & Bartolomeo, P. (2020). E-TAN, a technology-enhanced platform with tangible objects for the assessment of visual neglect: a multiple single-case study. Neuropsychological Rehabilitation, 1–15. 10.1080/09602011.2020.1762671. [DOI] [PubMed]
  51. Chiba Y, Nishihara K, Haga N. Evaluating visual bias and effect of proprioceptive feedback in unilateral neglect. Journal of Clinical Neuroscience. 2010;17(9):1148–1152. doi: 10.1016/j.jocn.2010.02.017. [DOI] [PubMed] [Google Scholar]
  52. Chicchi, G. I. A., Pallavicini, F., Pedroli, E., Serino, S., & Riva, G. (2015). Augmented Reality: A Brand New Challenge for the Assessment and Treatment of Psychological Disorders. Computational and Mathematical Methods in Medicine, 2015, 1–12. 10.1155/2015/862942 [DOI] [PMC free article] [PubMed]
  53. Chokron S, Bartolomeo P, Siéroff É. La négligence spatiale unilatérale: trente ans de recherches, de découvertes, d’espoirs et (surtout) de questions. Revue Neurologique. 2008;164:S134–S142. doi: 10.1016/S0035-3787(08)73304-6. [DOI] [PubMed] [Google Scholar]
  54. Chung SJ, Park E, Ye BS, Lee HS, Chang HJ, Song D, et al. The computerized table setting test for detecting unilateral neglect. PLOS ONE. 2016;11(1):e0147030. doi: 10.1371/journal.pone.0147030. [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Cocchini, G., & Beschin, N. (2020). The Fluff test: improved scoring system to account for different degrees of contralesional and ipsilesional personal neglect in brain damaged patients. Neuropsychological Rehabilitation, 1–15. 10.1080/09602011.2020.1797828. [DOI] [PubMed]
  56. Cocchini G, Beschin N, Jehkonen M. The Fluff Test: a simple task to assess body representation neglect. Neuropsychological Rehabilitation. 2001;11(1):17–31. doi: 10.1080/09602010042000132. [DOI] [Google Scholar]
  57. Colarusso, R. P., & Hammill, D. D. (1972). Motor-free visual perception test. Academic Therapy Pub.
  58. Corbetta M, Kincade MJ, Lewis C, Snyder AZ, Sapir A. Neural basis and recovery of spatial attention deficits in spatial neglect. Nature Neuroscience. 2005;8(11):1603–1610. doi: 10.1038/nn1574. [DOI] [PubMed] [Google Scholar]
  59. Corbetta M, Shulman GL. Spatial neglect and attention networks. Annual Review of Neuroscience. 2011;34(1):569–599. doi: 10.1146/annurev-neuro-061010-113731. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Corti C, Oprandi MC, Chevignard M, Jansari A, Oldrati V, Ferrari E, et al. Virtual-reality performance-based Assessment of cognitive functions in adult patients with acquired Brain Injury: a scoping review. Neuropsychology Review. 2021 doi: 10.1007/s11065-021-09498-0. [DOI] [PubMed] [Google Scholar]
  61. Czernuszenko A, Czlonkowska A. Risk factors for falls in stroke patients during inpatient rehabilitation. Clinical Rehabilitation. 2009;23(2):176–188. doi: 10.1177/0269215508098894. [DOI] [PubMed] [Google Scholar]
  62. Daini R, Angelelli P, Antonucci G, Cappa SF, Vallar G. Exploring the syndrome of spatial unilateral neglect through an illusion of length. Experimental Brain Research. 2002;144(2):224–237. doi: 10.1007/s00221-002-1034-8. [DOI] [PubMed] [Google Scholar]
  63. Dawson, A. M., Buxbaum, L. J., & Rizzo, A. A. (2008). The virtual reality lateralized attention test: Sensitivity and validity of a new clinical tool for assessing hemispatial neglect (pp.77–82). Presented at the 2008 Virtual Rehabilitation, IEEE.
  64. Delis, D. C., Kaplan, E., & Kramer, J. H. (2012). October 8). Delis-Kaplan executive function system. American Psychological Association. 10.1037/t15082-000.
  65. Della Sala S, Beschin N, Cubelli R. Persistent neglect in everyday life. Cortex; A Journal Devoted To The Study Of The Nervous System And Behavior. 2018;103:382–384. doi: 10.1016/j.cortex.2017.09.010. [DOI] [PubMed] [Google Scholar]
  66. Deouell, L. Y., Sacher, Y., & Soroker, N. (2005). Assessment of spatial attention after brain damage with a dynamic reaction time test. Journal of the International Neuropsychological Society, 11(06), 10.1017/S1355617705050824. [DOI] [PubMed]
  67. De Renzi, E. (1982). Disorders of space exploration and cognition. John Wiley.
  68. De Renzi E, Faglioni P, Scotti G. Hemispheric Contribution to Exploration of Space through the visual and tactile modality. Cortex; A Journal Devoted To The Study Of The Nervous System And Behavior. 1970;6(2):191–203. doi: 10.1016/S0010-9452(70)80027-2. [DOI] [PubMed] [Google Scholar]
  69. Diller L, Weinberg J. Hemi-inattention in rehabilitation. The evolution of a rational remediation program. Advances in Neurology. Hemi-Inattention and Hemisphere Specialization. 1977;18:63–82. [PubMed] [Google Scholar]
  70. Di Monaco M, Schintu S, Dotta M, Barba S, Tappero R, Gindri P. Severity of unilateral spatial neglect is an independent predictor of functional Outcome after Acute Inpatient Rehabilitation in individuals with Right Hemispheric Stroke. Archives of Physical Medicine and Rehabilitation. 2011;92(8):1250–1256. doi: 10.1016/j.apmr.2011.03.018. [DOI] [PubMed] [Google Scholar]
  71. Dvorkin AY, Bogey RA, Harvey RL, Patton JL. Mapping the neglected space: gradients of detection revealed by virtual reality. Neurorehabilitation and Neural Repair. 2012;26(2):120–131. doi: 10.1177/1545968311410068. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Ellis AW, Flude BM, Young AW. “Neglect dyslexia” and the early visual processing of letters in words and nonwords. Cognitive Neuropsychology. 1987;4(4):439–464. doi: 10.1080/02643298708252047. [DOI] [Google Scholar]
  73. Erez ABH, Katz N, Ring H, Soroker N. Assessment of spatial neglect using computerised feature and conjunction visual search tasks. Neuropsychological Rehabilitation. 2009;19(5):677–695. doi: 10.1080/09602010802711160. [DOI] [PubMed] [Google Scholar]
  74. Fels, M., & Geissner, E. (1997). Neglect-Test (NET): ein Verfahren zur Erfassung visueller Neglectphänomene; Handanweisung; deutsche überarbeitete Adaption des Behavioural Inattention Test (Wilson, Cockburn & Halligan, 1987). Hogrefe, Verlag für Psychologie.
  75. Ferbinteanu J. Memory systems 2018 – towards a new paradigm. Neurobiology of Learning and Memory. 2019;157:61–78. doi: 10.1016/j.nlm.2018.11.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Figueiredo L, Rodrigues E, Teixeira J, Teichrieb V. A comparative evaluation of direct hand and wand interactions on consumer devices. Computers & Graphics. 2018;77:108–121. doi: 10.1016/j.cag.2018.10.006. [DOI] [Google Scholar]
  77. Fordell H, Bodin K, Bucht G, Malm J. A virtual reality test battery for assessment and screening of spatial neglect: VR-test battery for assessment and screening of spatial neglect. Acta Neurologica Scandinavica. 2011;123(3):167–174. doi: 10.1111/j.1600-0404.2010.01390.x. [DOI] [PubMed] [Google Scholar]
  78. Gainotti G, Messerli P, Tissot R. Qualitative analysis of unilateral spatial neglect in relation to laterality of cerebral lesions. Journal of Neurology Neurosurgery & Psychiatry. 1972;35(4):545–550. doi: 10.1136/jnnp.35.4.545. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Galletta EE, Campanelli L, Maul KK, Barrett AM. Assessment of Neglect Dyslexia with Functional Reading materials. Topics in Stroke Rehabilitation. 2014;21(1):75–86. doi: 10.1310/tsr2101-75. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Gandola M, Toraldo A, Invernizzi P, Corrado L, Sberna M, Santilli I, et al. How many forms of perseveration? Evidence from cancellation tasks in right hemisphere patients. Neuropsychologia. 2013;51(14):2960–2975. doi: 10.1016/j.neuropsychologia.2013.10.023. [DOI] [PubMed] [Google Scholar]
  81. Gauthier L, Dehaut F, Joanette Y. The bells test: a quantitative and qualitative test for visual neglect. International journal of clinical neuropsychology. 1989;11(2):49–54. [Google Scholar]
  82. Geeraerts S, Lafosse C, Vandenbussche E, Verfaillie K. A psychophysical study of visual extinction: ipsilesional distractor interference with contralesional orientation thresholds in visual hemineglect patients. Neuropsychologia. 2005;43(4):530–541. doi: 10.1016/j.neuropsychologia.2004.07.012. [DOI] [PubMed] [Google Scholar]
  83. Geusgens CAV, Winkens I, van Heugten CM, Jolles J, van den Heuvel WJA. Occurrence and measurement of transfer in cognitive rehabilitation: a critical review. Journal of Rehabilitation Medicine. 2007;39(6):425–439. doi: 10.2340/16501977-0092. [DOI] [PubMed] [Google Scholar]
  84. Grattan ES, Woodbury ML. Do neglect assessments detect neglect differently? American Journal of Occupational Therapy. 2017;71(3):7103190050p1. doi: 10.5014/ajot.2017.025015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Grossi D, Modafferi A, Pelosi L, Trojano L. On the different roles of the cerebral hemispheres in mental imagery: the “o’Clock Test” in two clinical cases. Brain and Cognition. 1989;10(1):18–27. doi: 10.1016/0278-2626(89)90072-9. [DOI] [PubMed] [Google Scholar]
  86. Guariglia C, Padovani A, Pantano P, Pizzamiglio L. Unilateral neglect restricted to visual imagery. Nature. 1993;364(6434):235–237. doi: 10.1038/364235a0. [DOI] [PubMed] [Google Scholar]
  87. Halligan PW, Cockburn J, Wilson BA. The behavioural assessment of visual neglect. Neuropsychological Rehabilitation. 1991;1(1):5–32. doi: 10.1080/09602019108401377. [DOI] [Google Scholar]
  88. Halligan PW, Fink GR, Marshall JC, Vallar G. Spatial cognition: evidence from visual neglect. Trends in Cognitive Sciences. 2003;7(3):125–133. doi: 10.1016/S1364-6613(03)00032-9. [DOI] [PubMed] [Google Scholar]
  89. Halligan PW, Marshall JC, Wade DT. Visuospatial neglect: underlying factors and test sensitivity. The Lancet. 1989;334(8668):908–911. doi: 10.1016/S0140-6736(89)91561-4. [DOI] [PubMed] [Google Scholar]
  90. Heilman KM, Valenstein E, Watson RT. Neglect and related Disorders. Seminars in Neurology. 2000;20(04):463–470. doi: 10.1055/s-2000-13179. [DOI] [PubMed] [Google Scholar]
  91. Heilman, K. M., Watson, R. T., & Valenstein, E. (2003). Neglect and related disorders. [DOI] [PubMed]
  92. Holleman GA, Hooge ITC, Kemner C, Hessels RS. The ‘Real-World Approach’ and its problems: a critique of the term ecological validity. Frontiers in Psychology. 2020;11:721. doi: 10.3389/fpsyg.2020.00721. [DOI] [PMC free article] [PubMed] [Google Scholar]
  93. Hooper, E. H. (1983). Hooper visual organization test (VOT). Western Psychological Services.
  94. Howard, I. P. (1982). Human visual orientation. John Wiley & Sons.
  95. Ishiai S, Furukawa T, Tsukagoshi H. Visuospatial processes of line bisection and the mechanisms underlying unilateral spatia neglect. Brain. 1989;112(6):1485–1502. doi: 10.1093/brain/112.6.1485. [DOI] [PubMed] [Google Scholar]
  96. Isiai S. Behavioural inattention test, japanese edition. Tokyo, Japan: Shinkoh Igaku Shuppan, Co, Ltd; 1999. [Google Scholar]
  97. Jee H, Kim J, Kim C, Kim T, Park J. Feasibility of a semi-computerized line bisection test for Unilateral Visual Neglect Assessment. Applied Clinical Informatics. 2015;06(02):400–417. doi: 10.4338/ACI-2015-01-RA-0002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  98. Jewell G, McCourt ME. Pseudoneglect: a review and meta-analysis of performance factors in line bisection tasks. Neuropsychologia. 2000;38(1):93–110. doi: 10.1016/S0028-3932(99)00045-7. [DOI] [PubMed] [Google Scholar]
  99. Kan KC, Subramaniam P, Shahrizaila N, Kamaruzzaman SB, Razali R, Ghazali SE. Validation of the malay version of Addenbrooke’s cognitive examination III in detecting mild cognitive impairment and dementia. Dementia and Geriatric Cognitive Disorders Extra. 2019;9(1):66–76. doi: 10.1159/000495107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  100. Karnath HO, Rorden C. The anatomy of spatial neglect. Neuropsychologia. 2012;50(6):1010–1017. doi: 10.1016/j.neuropsychologia.2011.06.027. [DOI] [PMC free article] [PubMed] [Google Scholar]
  101. Kennedy RS, Lane NE, Berbaum KS, Lilienthal MG. Simulator Sickness Questionnaire: an enhanced method for quantifying Simulator sickness. The International Journal of Aviation Psychology. 1993;3(3):203–220. doi: 10.1207/s15327108ijap0303_3. [DOI] [Google Scholar]
  102. Kerkhoff G, Schenk T. Rehabilitation of neglect: an update. Neuropsychologia. 2012;50(6):1072–1079. doi: 10.1016/j.neuropsychologia.2012.01.024. [DOI] [PubMed] [Google Scholar]
  103. Kertesz, A. (1982). Western aphasia battery test manual. Psychological Corporation.
  104. Kühberger A, Fritz A, Scherndl T. Publication Bias in psychology: a diagnosis based on the correlation between Effect size and sample size. Plos One. 2014;9(9):e105825. doi: 10.1371/journal.pone.0105825. [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Kim DY, Ku J, Chang WH, Park TH, Lim JY, Han K, et al. Assessment of post-stroke extrapersonal neglect using a three-dimensional immersive virtual street crossing program. Acta Neurologica Scandinavica. 2010;121(3):171–177. doi: 10.1111/j.1600-0404.2009.01194.x. [DOI] [PubMed] [Google Scholar]
  106. Kim TL, Kim K, Choi C, Lee JY, Shin JH. FOPR test: a virtual reality-based technique to assess field of perception and field of regard in hemispatial neglect. Journal of NeuroEngineering and Rehabilitation. 2021;18(1):39. doi: 10.1186/s12984-021-00835-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  107. Kim YM, Chun MH, Yun GJ, Song YJ, Young HE. The effect of virtual reality training on unilateral spatial neglect in stroke patients. Annals of rehabilitation medicine. 2011;35(3):309. doi: 10.5535/arm.2011.35.3.309. [DOI] [PMC free article] [PubMed] [Google Scholar]
  108. Kinsbourne M, Warrington EK. A variety of reading disability associated with right hemisphere lesions. Journal of Neurology Neurosurgery and Psychiatry. 1962;25(4):339. doi: 10.1136/jnnp.25.4.339. [DOI] [PMC free article] [PubMed] [Google Scholar]
  109. Klatzky, R. L. (1998). Allocentric and egocentric spatial representations: definitions, distinctions, and interconnections (pp. 1–17). Presented at the Spatial cognition, Springer.
  110. Kleinman J, Dubois JC, Newhart M, Hillis AE. Disentangling the neuroanatomical correlates of perseveration from unilateral spatial neglect. Behavioural Neurology. 2013;26(1–2):131–138. doi: 10.1155/2013/438965. [DOI] [PMC free article] [PubMed] [Google Scholar]
  111. Knobel SEJ, Kaufmann BC, Gerber SM, Cazzoli D, Müri RM, Nyffeler T, Nef T. Immersive 3D virtual reality Cancellation Task for Visual Neglect Assessment: a pilot study. Frontiers in Human Neuroscience. 2020;14:180. doi: 10.3389/fnhum.2020.00180. [DOI] [PMC free article] [PubMed] [Google Scholar]
  112. Kocanaogullari, D., Mak, J., Kersey, J., Khalaf, A., Ostadabbas, S., Wittenberg, G. (2020). EEG-based Neglect Detection for Stroke Patients. In 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC) (pp. 264–267). Presented at the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC) in conjunction with the 43rd Annual Conference of the Canadian Medical and Biological Engineering Society, Montreal, QC, Canada: IEEE. 10.1109/EMBC44109.2020.9176378 [DOI] [PubMed]
  113. Kourtesis,., Korre D, Collina S, Doumas LAA, MacPherson SE. Guidelines for the development of immersive virtual reality Software for Cognitive Neuroscience and Neuropsychology: the development of virtual reality Everyday Assessment Lab (VR-EAL), a neuropsychological test battery in immersive virtual reality. Frontiers in Computer Science. 2020;1:12. doi: 10.3389/fcomp.2019.00012. [DOI] [Google Scholar]
  114. Kourtesis P, Collina S, Doumas LAA, MacPherson SE. Technological competence is a pre-condition for effective implementation of virtual reality head mounted Displays in Human Neuroscience: a Technological Review and Meta-Analysis. Frontiers in Human Neuroscience. 2019;13:342. doi: 10.3389/fnhum.2019.00342. [DOI] [PMC free article] [PubMed] [Google Scholar]
  115. Kourtesis P, Collina S, Doumas LAA, MacPherson SE. Validation of the virtual reality Neuroscience Questionnaire: Maximum Duration of Immersive virtual reality Sessions without the Presence of Pertinent adverse symptomatology. Frontiers in Human Neuroscience. 2019;13:417. doi: 10.3389/fnhum.2019.00417. [DOI] [PMC free article] [PubMed] [Google Scholar]
  116. Kourtesis P, MacPherson SE. How immersive virtual reality methods may meet the criteria of the National Academy of Neuropsychology and American Academy of Clinical Neuropsychology: A software review of the virtual reality Everyday Assessment Lab (VR-EAL) Computers in Human Behavior Reports. 2021;4:100151. doi: 10.1016/j.chbr.2021.100151. [DOI] [Google Scholar]
  117. Làdavas E. Selective spatial attention in patients with visual extinction. Brain. 1990;113(5):1527–1538. doi: 10.1093/brain/113.5.1527. [DOI] [PubMed] [Google Scholar]
  118. Lee BH, Kang SJ, Park JM, Son Y, Lee KH, Adair JC, et al. The character-line Bisection Task: a new test for hemispatial neglect. Neuropsychologia. 2004;42(12):1715–1724. doi: 10.1016/j.neuropsychologia.2004.02.015. [DOI] [PubMed] [Google Scholar]
  119. Levine DN, Warach JD, Benowitz L, Calvanio R. Left spatial neglect: Effects of lesion size and premorbid brain atrophy on severity and recovery following right cerebral infarction. Neurology. 1986;36(3):362–362. doi: 10.1212/WNL.36.3.362. [DOI] [PubMed] [Google Scholar]
  120. Lezak MD, Howieson DB, Bigler E, Tranel D. Neuropsychological assessment. 5. 9780195395525?cc=it〈=en&: Oxford University Press; 1995. [Google Scholar]
  121. Liang, Y., Guest, R. M., Fairhurst, M. C., & Potter, J. M. (2010). A computer-based quantitative assessment of visuo-spatial neglect using regression and data transformation.Pattern Analysis and Applications, 13(4),409–422. 10.1007/s10044-009-0172-z
  122. Machner B, Könemund I, von der Gablentz J, Bays PM, Sprenger A. The ipsilesional attention bias in right-hemisphere stroke patients as revealed by a realistic visual search task: neuroanatomical correlates and functional relevance. Neuropsychology. 2018;32(7):850–865. doi: 10.1037/neu0000493. [DOI] [PMC free article] [PubMed] [Google Scholar]
  123. Mainetti R, Sedda A, Ronchetti M, Bottini G, Borghese NA. Duckneglect: video-games based neglect rehabilitation. Technology and Health Care. 2013;21(2):97–111. doi: 10.3233/THC-120712. [DOI] [PubMed] [Google Scholar]
  124. Maister L, Slater M, Sanchez-Vives MV, Tsakiris M. Changing bodies changes minds: owning another body affects social cognition. Trends in Cognitive Sciences. 2015;19(1):6–12. doi: 10.1016/j.tics.2014.11.001. [DOI] [PubMed] [Google Scholar]
  125. Mancuso M, Rosadoni S, Capitani D, Bickerton WL, Humphreys GW, De Tanti A, et al. Italian standardization of the Apples Cancellation Test. Neurological Sciences. 2015;36(7):1233–1240. doi: 10.1007/s10072-015-2088-2. [DOI] [PubMed] [Google Scholar]
  126. Massironi M, Antonucci G, Pizzamiglio L, Vitale MV, Zoccolotti P. The Wundt-Jastrow illusion in the study of spatial hemi-inattention. Neuropsychologia. 1988;26(1):161–166. doi: 10.1016/0028-3932(88)90039-5. [DOI] [PubMed] [Google Scholar]
  127. McIntosh RD, Brodie EE, Beschin N, Robertson IH. Improving the clinical diagnosis of personal neglect: a reformulated comb and Razor Test. Cortex; A Journal Devoted To The Study Of The Nervous System And Behavior. 2000;36(2):289–292. doi: 10.1016/S0010-9452(08)70530-6. [DOI] [PubMed] [Google Scholar]
  128. Meador KJ, Loring DW, Bowers D, Heilman KM. Remote memory and neglect syndrome. Neurology. 1987;37(3):522–522. doi: 10.1212/WNL.37.3.522. [DOI] [PubMed] [Google Scholar]
  129. Menon A, Korner-Bitensky N. Evaluating unilateral spatial neglect Post Stroke: working your way through the maze of Assessment Choices. Topics in Stroke Rehabilitation. 2004;11(3):41–66. doi: 10.1310/KQWL-3HQL-4KNM-5F4U. [DOI] [PubMed] [Google Scholar]
  130. Mesa-Gresa, P., Lozano, J. A., Llórens, R., Alcañiz, M., Navarro, M. D., & Noé, E. (2011). Clinical validation of a virtual environment test for safe street crossing in the assessment of acquired brain injury patients with and without neglect (pp. 44–51). Presented at the IFIP Conference on Human-Computer Interaction, Springer.
  131. Milner AD, Harvey M, Roberts RC, Forster SV. Line bisection errors in visual neglect: misguided action or size distortion? Neuropsychologia. 1993;31(1):39–49. doi: 10.1016/0028-3932(93)90079-F. [DOI] [PubMed] [Google Scholar]
  132. Mizuno K, Kato K, Tsuji T, Shindo K, Kobayashi Y, Liu M. Spatial and temporal dynamics of visual search tasks distinguish subtypes of unilateral spatial neglect: comparison of two cases with viewer-centered and stimulus-centered neglect. Neuropsychological Rehabilitation. 2016;26(4):610–634. doi: 10.1080/09602011.2015.1051547. [DOI] [PubMed] [Google Scholar]
  133. Montedoro V, Alsamour M, Dehem S, Lejeune T, Dehez B, Edwards MG. Robot diagnosis test for egocentric and allocentric hemineglect. Archives of Clinical Neuropsychology. 2019;34(4):481–494. doi: 10.1093/arclin/acy062. [DOI] [PubMed] [Google Scholar]
  134. Munn, Z., Peters, M. D., Stern, C., Tufanaru, C., McArthur, A., & Aromataris, E. (2018). Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review approach. BMC Medical Research Methodology, 18, 1–7. [DOI] [PMC free article] [PubMed]
  135. Murphy TH, Corbett D. Plasticity during stroke recovery: from synapse to behaviour. Nature Reviews Neuroscience. 2009;10(12):861–872. doi: 10.1038/nrn2735. [DOI] [PubMed] [Google Scholar]
  136. Neguț A, Matu SA, Sava FA, David D. Virtual reality measures in neuropsychological assessment: a meta-analytic review. The Clinical Neuropsychologist. 2016;30(2):165–184. doi: 10.1080/13854046.2016.1144793. [DOI] [PubMed] [Google Scholar]
  137. Niedeggen, M., & Jörgens, S. (2005). Visuelle Wahrnehmungsstörungen (5 vol.). Hogrefe Verlag.
  138. Nijboer, T., van de Port, I., Schepers, V., Post, M., & Visser-Meily, A. (2013). Predicting Functional Outcome after Stroke: the influence of neglect on Basic Activities in Daily Living. Frontiers in Human Neuroscience, 7. 10.3389/fnhum.2013.00182. [DOI] [PMC free article] [PubMed]
  139. Ogden JA. Contralesional neglect of constructed visual images in right and left brain-damaged patients. Neuropsychologia. 1985;23(2):273–277. doi: 10.1016/0028-3932(85)90112-5. [DOI] [PubMed] [Google Scholar]
  140. Ogourtsova T, Archambault PS, Lamontagne A. Post-stroke unilateral spatial neglect: virtual reality-based navigation and detection tasks reveal lateralized and non-lateralized deficits in tasks of varying perceptual and cognitive demands. Journal of NeuroEngineering and Rehabilitation. 2018;15(1):34. doi: 10.1186/s12984-018-0374-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  141. Ogourtsova T, Archambault PS, Sangani S, Lamontagne A. Ecological virtual reality evaluation of neglect symptoms (EVENS): Effects of virtual scene complexity in the Assessment of Poststroke unilateral spatial neglect. Neurorehabilitation and Neural Repair. 2018;32(1):46–61. doi: 10.1177/1545968317751677. [DOI] [PubMed] [Google Scholar]
  142. Ogourtsova T, Souza Silva W, Archambault PS, Lamontagne A. Virtual reality treatment and assessments for post-stroke unilateral spatial neglect: a systematic literature review. Neuropsychological Rehabilitation. 2017;27(3):409–454. doi: 10.1080/09602011.2015.1113187. [DOI] [PubMed] [Google Scholar]
  143. Ota H, Fujii T, Suzuki K, Fukatsu R, Yamadori A. Dissociation of body-centered and stimulus-centered representations in unilateral neglect. Neurology. 2001;57(11):2064–2069. doi: 10.1212/WNL.57.11.2064. [DOI] [PubMed] [Google Scholar]
  144. Ota H, Fujii T, Tabuchi M, Sato K, Saito J, Yamadori A. Different spatial processing for stimulus-centered and body-centered representations. Neurology. 2003;60(11):1846–1848. doi: 10.1212/01.WNL.0000065899.22903.07. [DOI] [PubMed] [Google Scholar]
  145. Ouzzani M, Hammady H, Fedorowicz Z, Elmagarmid A. Rayyan—a web and mobile app for systematic reviews. Systematic Reviews. 2016;5(1):210. doi: 10.1186/s13643-016-0384-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  146. Pallavicini F, Pedroli E, Serino S, Dell’Isola A, Cipresso P, Cisari C, Riva G. Assessing unilateral spatial neglect using advanced technologies: the potentiality of mobile virtual reality. Technology and Health Care. 2015;23(6):795–807. doi: 10.3233/THC-151039. [DOI] [PubMed] [Google Scholar]
  147. Parsons, T. D. (2015). Virtual reality for enhanced ecological validity and experimental control in the Clinical, Affective and Social Neurosciences. Frontiers in Human Neuroscience, 9. 10.3389/fnhum.2015.00660. [DOI] [PMC free article] [PubMed]
  148. Paterson A, Zangwill OL. Disorders of visual space perception associated with lesions of the right cerebral hemisphere. Brain. 1944;67(4):331–358. doi: 10.1093/brain/67.4.331. [DOI] [Google Scholar]
  149. Pedroli, E., Serino, S., Cipresso, P., Pallavicini, F., & Riva, G. (2015). Assessment and rehabilitation of neglect using virtual reality: a systematic review. Frontiers in Behavioral Neuroscience, 9. 10.3389/fnbeh.2015.00226. [DOI] [PMC free article] [PubMed]
  150. Peru A, Bollini A, Costanzo S, Dainelli M, Landucci L, Pezzatini D, et al. Towards a dynamic Assessment of unilateral spatial neglect. The Open Psychology Journal. 2017;10(1):71–80. doi: 10.2174/1874350101710010071. [DOI] [Google Scholar]
  151. Peskine A, Rosso C, Box N, Galland A, Caron E, Rautureau G, et al. Virtual reality assessment for visuospatial neglect: importance of a dynamic task. Journal of Neurology Neurosurgery & Psychiatry. 2011;82(12):1407–1409. doi: 10.1136/jnnp.2010.217513. [DOI] [PubMed] [Google Scholar]
  152. Pierce, J. E., Ronchi, R., Thomasson, M., Rossi, I., Casati, C., Saj, A., et al. (2021). A novel computerized assessment of manual spatial exploration in unilateral spatial neglect. Neuropsychological Rehabilitation, 1–22. 10.1080/09602011.2021.1875850. [DOI] [PubMed]
  153. Pizzamiglio, L., Judica, A., Razzano, C., & Zoccolotti, P. (1989). Toward a comprehensive diagnosis of visual-spatial disorders in unilateral brain damaged patients. Evaluación Psicológica.
  154. Ptak R, Schnider A. The attention network of the human brain: relating structural damage associated with spatial neglect to functional imaging correlates of spatial attention. Neuropsychologia. 2011;49(11):3063–3070. doi: 10.1016/j.neuropsychologia.2011.07.008. [DOI] [PubMed] [Google Scholar]
  155. Quinn TJ, Livingstone I, Weir A, Shaw R, Breckenridge A, McAlpine C, Tarbert CM. Accuracy and feasibility of an android-based Digital Assessment Tool for Post Stroke Visual Disorders—The StrokeVision App. Frontiers in Neurology. 2018;9:146. doi: 10.3389/fneur.2018.00146. [DOI] [PMC free article] [PubMed] [Google Scholar]
  156. Rabuffetti M, Farina E, Alberoni M, Pellegatta D, Appollonio I, Affanni P, et al. Spatio-temporal features of Visual Exploration in unilaterally brain-damaged subjects with or without neglect: results from a touchscreen test. Plos One. 2012;7(2):e31511. doi: 10.1371/journal.pone.0031511. [DOI] [PMC free article] [PubMed] [Google Scholar]
  157. Rabuffetti M, Ferrarin M, Spadone R, Pellegatta D, Gentileschi V, Vallar G, Pedotti A. Touch-screen system for assessing visuo-motor exploratory skills in neuropsychological disorders of spatial cognition. Medical & Biological Engineering & Computing. 2002;40(6):675–686. doi: 10.1007/BF02345306. [DOI] [PubMed] [Google Scholar]
  158. Rand D, Kizony R, Weiss PTL. The Sony PlayStation II EyeToy: low-cost virtual reality for use in rehabilitation. Journal of neurologic physical therapy. 2008;32(4):155–163. doi: 10.1097/NPT.0b013e31818ee779. [DOI] [PubMed] [Google Scholar]
  159. Rehme AK, Eickhoff SB, Rottschy C, Fink GR, Grefkes C. Activation likelihood estimation meta-analysis of motor-related neural activity after stroke. Neuroimage. 2012;59(3):2771–2782. doi: 10.1016/j.neuroimage.2011.10.023. [DOI] [PubMed] [Google Scholar]
  160. Reinhart S, Leonhard E, Kerkhoff G. Sensitive neglect test (SNT) single and dual task. Saarbrücken: Saarland University; 2016. [Google Scholar]
  161. Ringman JM, Saver JL, Woolson RF, Clarke WR, Adams HP. Frequency, risk factors, anatomy, and course of unilateral neglect in an acute stroke cohort. Neurology. 2004;63(3):468–474. doi: 10.1212/01.WNL.0000133011.10689.CE. [DOI] [PubMed] [Google Scholar]
  162. Rizzo AA, Schultheis M, Kerns KA, Mateer C. Analysis of assets for virtual reality applications in neuropsychology. Neuropsychological Rehabilitation. 2004;14(1–2):207–239. doi: 10.1080/09602010343000183. [DOI] [Google Scholar]
  163. Robey RR, Schultz MC, Crawford AB, Sinner CA. Single-subject clinical-outcome research: designs, data, effect sizes, and analyses. Aphasiology. 1999;13(6):445–473. doi: 10.1080/026870399402028. [DOI] [Google Scholar]
  164. Rode G, Michel C, Rossetti Y, Boisson D, Vallar G. Left size distortion (hyperschematia) after right brain damage. Neurology. 2006;67(10):1801–1808. doi: 10.1212/01.wnl.0000244432.91915.d0. [DOI] [PubMed] [Google Scholar]
  165. Rode G, Perenin MT, Boisson D. Neglect of the representational space: demonstration by mental evocation of the map of France. Revue neurologique. 1995;151(3):161–164. [PubMed] [Google Scholar]
  166. Ronchi R, Algeri L, Chiapella L, Gallucci M, Spada MS, Vallar G. Left neglect dyslexia: perseveration and reading error types. Neuropsychologia. 2016;89:453–464. doi: 10.1016/j.neuropsychologia.2016.07.023. [DOI] [PubMed] [Google Scholar]
  167. Ronchi R, Algeri L, Chiapella L, Spada MS, Vallar G. Spatial neglect and perseveration in visuomotor exploration. Neuropsychology. 2012;26(5):588–603. doi: 10.1037/a0029216. [DOI] [PubMed] [Google Scholar]
  168. Ronchi R, Posteraro L, Fortis P, Bricolo E, Vallar G. Perseveration in left spatial neglect: drawing and cancellation tasks. Cortex; A Journal Devoted To The Study Of The Nervous System And Behavior. 2009;45(3):300–312. doi: 10.1016/j.cortex.2008.03.012. [DOI] [PubMed] [Google Scholar]
  169. Rorden C, Karnath HO. A simple measure of neglect severity. Neuropsychologia. 2010;48(9):2758–2763. doi: 10.1016/j.neuropsychologia.2010.04.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  170. Rossa, T., Pompidor, P., Rodriguez, N., Sallaberry, A., Poncelet, P., Urbanski, M. (2019). Experimental web service and eye-tracking setup for unilateral spatial neglect assessment (pp. 141–155). Presented at the International Conference on Human-Computer Interaction, Springer.
  171. Rusconi ML, Maravita A, Bottini G, Vallar G. Is the intact side really intact? Perseverative responses in patients with unilateral neglect: a productive manifestation. Neuropsychologia. 2002;40(6):594–604. doi: 10.1016/S0028-3932(01)00160-9. [DOI] [PubMed] [Google Scholar]
  172. Schendel KL, Robertson LC. Using reaction time to assess patients with unilateral neglect and extinction. Journal of Clinical and Experimental Neuropsychology. 2002;24(7):941–950. doi: 10.1076/jcen.24.7.941.8390. [DOI] [PubMed] [Google Scholar]
  173. Schenkenberg T, Bradford DC, Ajax ET. Line bisection and unilateral visual neglect in patients with neurologic impairment. Neurology. 1980;30(5):509–509. doi: 10.1212/WNL.30.5.509. [DOI] [PubMed] [Google Scholar]
  174. Schwartz MF, Segal M, Veramonti T, Ferraro M, Buxbaum LJ. The naturalistic action test: a standardised assessment for everyday action impairment. Neuropsychological Rehabilitation. 2002;12(4):311–339. doi: 10.1080/09602010244000084. [DOI] [Google Scholar]
  175. Sedda A, Borghese NA, Ronchetti M, Mainetti R, Pasotti F, Beretta G, Bottini G. Using virtual reality to rehabilitate neglect. Behavioural Neurology. 2013;26(3):183–185. doi: 10.1155/2013/810279. [DOI] [PMC free article] [PubMed] [Google Scholar]
  176. Siéroff E. Acquired spatial dyslexia. Annals of Physical and Rehabilitation Medicine. 2017;60(3):155–159. doi: 10.1016/j.rehab.2015.07.004. [DOI] [PubMed] [Google Scholar]
  177. Slater M. Place illusion and plausibility can lead to realistic behaviour in immersive virtual environments. Philosophical Transactions of the Royal Society B: Biological Sciences. 2009;364(1535):3549–3557. doi: 10.1098/rstb.2009.0138. [DOI] [PMC free article] [PubMed] [Google Scholar]
  178. Slater M. Immersion and the illusion of presence in virtual reality. British Journal of Psychology. 2018;109(3):431–433. doi: 10.1111/bjop.12305. [DOI] [PubMed] [Google Scholar]
  179. Smit, M., Van der Stigchel, S., Visser-Meily, J. M. A., Kouwenhoven, M., Eijsackers, A. L. H., & Nijboer, T. C. W. (2013). The feasibility of computer-based prism adaptation to ameliorate neglect in Sub-Acute Stroke Patients admitted to a Rehabilitation Center. Frontiers in Human Neuroscience, 7. 10.3389/fnhum.2013.00353. [DOI] [PMC free article] [PubMed]
  180. Sotokawa T, Murayama T, Noguchi J, Sakimura Y, Itoh M. Driving evaluation of mild unilateral spatial neglect patients–three high-risk cases undetected by BIT after Recovery. In: Yamamoto S, editor. Human interface and the management of information. Information and knowledge in Context. Cham: Springer International Publishing; 2015. pp. 253–261. [Google Scholar]
  181. Sportillo, D., Paljic, A., Boukhris, M., Fuchs, P., Ojeda, L., & Roussarie, V. (2017). An Immersive Virtual Reality System for Semi-autonomous Driving Simulation: A Comparison between Realistic and 6-DoF Controller-based Interaction. In Proceedings of the 9th International Conference on Computer and Automation Engineering - ICCAE ’17 (pp. 6–10). Presented at the the 9th International Conference, Sydney, Australia: ACM Press. 10.1145/3057039.3057079
  182. Spreij LA, Brink T, Visser-Meily AF, Nijboer TCW. Simulated driving: the added value of dynamic testing in the assessment of visuo‐spatial neglect after stroke. Journal of Neuropsychology. 2020;14(1):28–45. doi: 10.1111/jnp.12172. [DOI] [PMC free article] [PubMed] [Google Scholar]
  183. Stefaniak JD, Halai AD, Lambon Ralph MA. The neural and neurocomputational bases of recovery from post-stroke aphasia. Nature Reviews Neurology. 2020;16(1):43–55. doi: 10.1038/s41582-019-0282-1. [DOI] [PubMed] [Google Scholar]
  184. Stinear CM. Prediction of motor recovery after stroke: advances in biomarkers. The Lancet Neurology. 2017;16(10):826–836. doi: 10.1016/S1474-4422(17)30283-1. [DOI] [PubMed] [Google Scholar]
  185. Stone SP, Wilson B, Wroot A, Halligan PW, Lange LS, Marshall JC, Greenwood RJ. The assessment of visuo-spatial neglect after acute stroke. Journal of Neurology Neurosurgery & Psychiatry. 1991;54(4):345–350. doi: 10.1136/jnnp.54.4.345. [DOI] [PMC free article] [PubMed] [Google Scholar]
  186. Strauss ME, Smith GT. Construct validity: advances in theory and methodology. Annual Review of Clinical Psychology. 2009;5(1):1–25. doi: 10.1146/annurev.clinpsy.032408.153639. [DOI] [PMC free article] [PubMed] [Google Scholar]
  187. Sugarman, H., Weisel-Eichler, A., Burstin, A., & Brown, R. (2011). Use of novel virtual reality system for the assessment and treatment of unilateral spatial neglect: A feasibility study (pp. 1–2). Presented at the 2011 International Conference on Virtual Rehabilitation, IEEE.
  188. Swift A, Heale R, Twycross A. What are sensitivity and specificity? Evidence Based Nursing. 2020;23(1):2–4. doi: 10.1136/ebnurs-2019-103225. [DOI] [PubMed] [Google Scholar]
  189. Tanaka T. Clinical application for assistive engineering-mixed reality rehabilitation. Journal of Medical and Biological Engineering. 2010;31(4):277–282. doi: 10.5405/jmbe.792. [DOI] [Google Scholar]
  190. Teuber HL. Physiological psychology. Annual review of psychology. 1955;6(1):267–296. doi: 10.1146/annurev.ps.06.020155.001411. [DOI] [PubMed] [Google Scholar]
  191. Tham K. The Baking Tray Task: a test of spatial neglect. Neuropsychological Rehabilitation. 1996;6(1):19–26. doi: 10.1080/713755496. [DOI] [PubMed] [Google Scholar]
  192. Toraldo A, Romaniello C, Sommaruga P. Measuring and diagnosing unilateral neglect: a standardized statistical procedure. The Clinical Neuropsychologist. 2017;31(6–7):1248–1267. doi: 10.1080/13854046.2017.1349181. [DOI] [PubMed] [Google Scholar]
  193. Tricco AC, Lillie E, Zarin W, O’Brien KK, Colquhoun H, Levac D, et al. PRISMA Extension for scoping reviews (PRISMA-ScR): Checklist and Explanation. Annals of Internal Medicine. 2018;169(7):467–473. doi: 10.7326/M18-0850. [DOI] [PubMed] [Google Scholar]
  194. Tsirlin I, Dupierrix E, Chokron S, Coquillart S, Ohlmann T. Uses of virtual reality for diagnosis, Rehabilitation and Study of unilateral spatial neglect: review and analysis. CyberPsychology & Behavior. 2009;12(2):175–181. doi: 10.1089/cpb.2008.0208. [DOI] [PubMed] [Google Scholar]
  195. Ulm L, Wohlrapp D, Meinzer M, Steinicke R, Schatz A, Denzler P, et al. A Circle-Monitor for Computerised Assessment of Visual Neglect in Peripersonal Space. Plos One. 2013;8(12):e82892. doi: 10.1371/journal.pone.0082892. [DOI] [PMC free article] [PubMed] [Google Scholar]
  196. Vaes N, Lafosse C, Nys G, Schevernels H, Dereymaeker L, Oostra K, et al. Capturing peripersonal spatial neglect: an electronic method to quantify visuospatial processes. Behavior Research Methods. 2015;47(1):27–44. doi: 10.3758/s13428-014-0448-0. [DOI] [PubMed] [Google Scholar]
  197. Vallar G. Spatial hemineglect in humans. Trends in Cognitive Sciences. 1998;2(3):87–97. doi: 10.1016/S1364-6613(98)01145-0. [DOI] [PubMed] [Google Scholar]
  198. Vallar G. The methodological foundations of human neuropsychology; studies in brain damaged patients. In: Boller F, Grafman J, Rizzolatti G, editors. Handbook of Neuropsychology. Amsterdam, The Netherlands: Elsevier; 2000. pp. 305–344. [Google Scholar]
  199. Vallar, G., & Bolognini, N. (2014). Unilateral Spatial Neglect. (A. C. Nobre & S. Kastner, Eds.) (Vol. 1). Oxford University Press. 10.1093/oxfordhb/9780199675111.013.012
  200. Vallar G, Bottini G, Sterzi R, Passerini D, Rusconi ML. Hemianesthesia, sensory neglect, and defective access to conscious experience. Neurology. 1991;41(5):650–652. doi: 10.1212/WNL.41.5.650. [DOI] [PubMed] [Google Scholar]
  201. Vallar G, Burani C, Arduino LS. Neglect dyslexia: a review of the neuropsychological literature. Experimental Brain Research. 2010;206(2):219–235. doi: 10.1007/s00221-010-2386-0. [DOI] [PubMed] [Google Scholar]
  202. Vallar, G., & Calzolari, E. (2018). Unilateral spatial neglect after posterior parietal damage. In Handbook of Clinical Neurology (Vol. 151, pp. 287–312). Elsevier. 10.1016/B978-0-444-63622-5.00014-0 [DOI] [PubMed]
  203. Vallar, G., & Caputi, N. (2021). The History of Human Neuropsychology. In S. Della Sala (Ed.), Encyclopedia of Behavioral Neuroscience (Vol. 1, pp. 14–39). Elsevier. 10.1016/b978-0-12-809324-5.23914-x
  204. Vallar G, Daini R, Antonucci G. Processing of illusion of length in spatial hemineglect: a study of line bisection. Neuropsychologia. 2000;38(7):1087–1097. doi: 10.1016/S0028-3932(99)00139-6. [DOI] [PubMed] [Google Scholar]
  205. Vallar G, Maravita A. Personal and extrapersonal spatial perception. In: Berntson GG, Cacioppo JT, editors. Handbook of Neuroscience for the behavioral Sciences (p. neubb001016) Hoboken, NJ, USA: John Wiley & Sons, Inc; 2009. [Google Scholar]
  206. Vallar G, Perani D. The anatomy of unilateral neglect after right-hemisphere stroke lesions. A clinical/CT-scan correlation study in man. Neuropsychologia. 1986;24(5):609–622. doi: 10.1016/0028-3932(86)90001-1. [DOI] [PubMed] [Google Scholar]
  207. Vallar G, Ronchi R. Somatoparaphrenia: a body delusion. A review of the neuropsychological literature. Experimental Brain Research. 2009;192(3):533–551. doi: 10.1007/s00221-008-1562-y. [DOI] [PubMed] [Google Scholar]
  208. Vallar, G., Rusconi, M. L., Fontana, S., & Musicco, M. (1994). Tre test di esplorazione visuo-spaziale: taratura su 212 soggetti normali. Archivio di Psicologia, Neurologia e Psichiatria.
  209. Vallar G, Sandroni P, Rusconi ML, Barbieri S. Hemianopia, hemianesthesia, and spatial neglect: a study with evoked potentials. Neurology. 1991;41(12):1918–1918. doi: 10.1212/WNL.41.12.1918. [DOI] [PubMed] [Google Scholar]
  210. Van der Stigchel S, Nijboer TCW. The imbalance of oculomotor capture in unilateral visual neglect. Consciousness and Cognition. 2010;19(1):186–197. doi: 10.1016/j.concog.2009.11.003. [DOI] [PubMed] [Google Scholar]
  211. Van der Stoep N, Visser-Meily JMA, Kappelle LJ, de Kort PLM, Huisman KD, Eijsackers ALH, et al. Exploring near and far regions of space: Distance-specific visuospatial neglect after stroke. Journal of Clinical and Experimental Neuropsychology. 2013;35(8):799–811. doi: 10.1080/13803395.2013.824555. [DOI] [PubMed] [Google Scholar]
  212. Villarreal, S., Linnavuo, M., Sepponen, R., Vuori, O., Bonato, M., Jokinen, H., & Hietanen, M. (2021). Unilateral stroke: computer-based Assessment uncovers non-lateralized and Contralesional Visuoattentive deficits. Journal of the International Neuropsychological Society, 1–11. 10.1017/S1355617720001393. [DOI] [PubMed]
  213. Villarreal S, Linnavuo M, Sepponen R, Vuori O, Bonato M, Jokinen H, Hietanen M. Computer-based Assessment: Dual-Task outperforms large-screen Cancellation Task in detecting Contralesional Omissions. Frontiers in Psychology. 2022;12:790438. doi: 10.3389/fpsyg.2021.790438. [DOI] [PMC free article] [PubMed] [Google Scholar]
  214. Villarreal S, Linnavuo M, Sepponen R, Vuori O, Jokinen H, Hietanen M. Dual-Task in large Perceptual Space reveals subclinical hemispatial neglect. Journal of the International Neuropsychological Society. 2020;26(10):993–1005. doi: 10.1017/S1355617720000508. [DOI] [PubMed] [Google Scholar]
  215. Walker R. Spatial and object-based neglect. Neurocase. 1995;1(4):371–383. doi: 10.1080/13554799508402381. [DOI] [Google Scholar]
  216. Wechsler, D. (2009). WMS-IV: Wechsler memory scale. Pearson.
  217. Wechsler, D. (2012). November 12). Wechsler Adult Intelligence Scale–Fourth Edition. American Psychological Association. 10.1037/t15169-000.
  218. Weintraub S, Mesulam MM. Visual hemispatial inattention: stimulus parameters and exploratory strategies. Journal of Neurology Neurosurgery & Psychiatry. 1988;51(12):1481–1488. doi: 10.1136/jnnp.51.12.1481. [DOI] [PMC free article] [PubMed] [Google Scholar]
  219. Weisenberg, T., & McBride, K. E. (1935). Aphasia: a clinical and psychological study. Aphasia: a clinical and psychological study. The Commonwealth Fund.
  220. Willer L, Pedersen PM, Forchhammer HB, Christensen H. Cognitive assessment at bedside for iPad: a preliminary validation of a novel cognitive test for stroke patients. European Stroke Journal. 2016;1(4):294–301. doi: 10.1177/2396987316665233. [DOI] [PMC free article] [PubMed] [Google Scholar]
  221. Wilson, B., Cockburn, J., & Halligan, P. W. (1987). Behavioural inattention test thames valley test company: Titchfield. Hampshire, UK.
  222. Yesavage JA, Sheikh JI. 9/Geriatric Depression Scale (GDS): recent evidence and development of a shorter version. Clinical Gerontologist. 1986;5(1–2):165–173. doi: 10.1300/J018v05n01_09. [DOI] [Google Scholar]
  223. Zhang J, Yu J, Bao Y, Xie Q, Xu Y, Zhang J, Wang P. Constraint-induced aphasia therapy in post-stroke aphasia rehabilitation: a systematic review and meta-analysis of randomized controlled trials. PLOS ONE. 2017;12(8):e0183349. doi: 10.1371/journal.pone.0183349. [DOI] [PMC free article] [PubMed] [Google Scholar]
  224. Zimmermann, P., & Fimm, B. (2002). In P. Z. M. Leclercq (Ed.), A test battery for attentional performance. Theory, diagnosis and rehabilitation.
  225. Zoccolotti P, Antonucci G, Judica A. Psychometric characteristics of two semi-structured scales for the functional evaluation of hemi-inattention in extrapersonal and personal space. Neuropsychological Rehabilitation. 1992;2(3):179–191. doi: 10.1080/09602019208401407. [DOI] [Google Scholar]
  226. Zoccolotti P, Antonucci G, Judica A, Montenero P, Pizzamiglio L, Razzano C. Incidence and evolution of the Hemineglect disorder in chronic patients with unilateral right brain damage. International Journal of Neuroscience. 1989;47(3–4):209–216. doi: 10.3109/00207458908987435. [DOI] [PubMed] [Google Scholar]
  227. Zoccolotti P, Judica A. Functional evaluation of hemineglect by means of a semistructured scale: personal extrapersonal differentiation. Neuropsychological Rehabilitation. 1991;1(1):33–44. doi: 10.1080/09602019108401378. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

data sharing not applicable to this article as no datasets were generated or analyzed during the current study.


Articles from Neuropsychology Review are provided here courtesy of Springer

RESOURCES