Abstract
Visual prostheses represent a groundbreaking avenue for restoring vision in individuals with visual impairments. These devices utilize electrode arrays positioned in early visual processing areas, like the retina, thalamus, or primary visual cortex. Connected to a camera, they transform a stream of video to electrical stimulation to present the visual environment through patterned activation of phosphenes. Visual prostheses offer the potential to enhance visual function and thereby quality of life for users, however, understanding and replicating motion perception in a manner akin to natural vision remains a critical challenge for device designers. This review presents studies of motion perception in different visual prosthesis modalities and discusses their advantages and limitations. Retinal and cortical visual prostheses show significant potential in enhancing motion perception, but many implementations have shortcomings. Some challenges which remain for better motion perception in visual prosthesis are gaze contingency, the effective integration of machine vision, and understanding the involvement of higher-order visual areas. Despite these challenges, the current research should be viewed with substantial optimism for the future of restoring functional vision to visually impaired individuals.
Keywords: artificial vision, phosphenes, frontoparallel motion, looming detection, gaze contingency
1. Introduction
Visual impairment is a pervasive global issue affecting millions of individuals (Bourne et al 2020). Those living with visual impairment experience profound and multifaceted impacts on their physical (Krumpaszky and Klauss 1996), cognitive (Zheng et al 2018, Lim et al 2020), emotional (Evans et al 2007), psycho-social (Salminen and Karhula 2014, Swenor et al 2020), and financial well-being (Köberlein et al 2013). Furthermore, visual impairment has substantial socioeconomic consequences (Eckert et al 2022, Rein et al 2022). As growing and aging populations make the global burden of blindness and visual impairment a pressing issue, exploring visual prosthesis devices as a treatment has shown considerable promise. With many visual prosthesis devices on the market (summarized in Makawi et al 2024) and others in clinical trials, this review aims to evaluate the current state of visual prostheses with regard to the perception of motion they can offer.
As of this writing, many different visual prosthesis systems have either been approved for clinical use, are currently in, or are nearing human clinical trials. These include the Alpha IMS/AMS, Argus II, CORTIVIS, EPI-RET, ICVP, IRIS, Orion, PRIMA, and others (summarized in Makawi et al 2024). However, while some of these systems are in active development, some have been abandoned or are no longer available; the implications for implanted patients have been discussed in the press (Strickland and Harris 2022) and scientific literature (e.g. Ramirez et al 2023). While much of the existing clinical work has concentrated on the retina, other labs are developing prosthetic systems to stimulate visual areas outside of the retina (reviewed in Mirochnik and Pezaris 2019), but with a few exceptions, the majority of this later work remains in pre-clinical testing.
In this review, we will summarize the current body of knowledge concerning motion perception in artificial vision and outline the importance of concentrating on this aspect of prosthesis function. Starting with a review of the neurological basis of motion perception, we will cover perception of frontoparallel motion, the most common paradigm used to assess motion processing in artificial vision. Tests of motion perception have been studied in visual prostheses that stimulate several different anatomical locations, including the retina, the visual cortex, and area MT. In addition to visual prosthesis devices, we will review the current literature on motion perception in artificial vision simulations. Finally, we will discuss the broader implications and challenges for achieving reliable motion perception in prosthetic vision. A summary of the systems considered during this review can be found in table 1, along with additional themes and concerns in table 2.
Table 1.
Comparison of prosthesis approaches and motion perception. This table compares major classes of visual prostheses based on implantation site, mechanism of action, clinical or experimental status, and their implications for motion perception. For each approach, the table lists key advantages and limitations, representative devices or descriptive publications, estimated visual field and resolution, and whether motion perception has been studied in published work. This summary shows the range of engineering strategies currently under investigation and the varying degrees of empirical support for motion perception across approaches.
| Approach | Mechanism | Advantages | Limitations | Examples | Visual field | Resolution | Motion studies |
|---|---|---|---|---|---|---|---|
| Epiretinal | Electrical stimulation of ganglion cells from the inner retinal surface | Established clinical use; direct access to output neurons | Requires intact inner retina; head-fixed camera disrupts gaze linkage | Argus II (Second Sight; discontinued) | ∼20° | 60 electrodes | Dorn et al (2013), Ho et al (2015) |
|
| |||||||
| Subretinal | Photodiode array implanted beneath the retina to stimulate bipolar cells directly | Naturalistic input by imaging external world; eye-movement integration | Requires functional bipolar cells; image quality depends on light conditions | Alpha AMS/IMS (Retina Implant AG; discontinued), PRIMA (science corp) | ∼15° | Alpha AMS: 1600 microphotodiodes; PRIMA: 378 pixels | Stingl et al (2015), Edwards et al (2018) |
|
| |||||||
| Suprachoroidal | Electrode array placed between sclera and choroid, stimulating outer retina | Safer, less invasive surgery; good biocompatibility | Greater distance from target neurons reduces spatial resolution | Bionic vision Australia (BVA) prototype, Monash Vision Group | ∼10°–30° | 44–98 electrodes | Titchener et al (2020) |
|
| |||||||
| Thalamic | Stimulation of the lateral geniculate nucleus (LGN) | Applicable for patients with retinal/optic nerve damage | Less explored clinically | Research prototypes; Pezaris and Reid (2007); Phosphoenix, BV | Potentially full field | Experimental (1000–2000) | None |
|
| |||||||
| Cortical | Stimulation of visual cortex (typically V1 or extrastriate areas) | Applicable for patients with retinal/optic nerve damage | Poor motion selectivity; limited spatial integration | Orion (Cortigent) ICVP (IIT) CORTIVIS (UMH) | ∼10°–20° (modular islands) | Orion: 60 electrodes; ICVP: 256 (modular); CORTIVIS: 100 (modular) | Beauchamp et al (2020) |
|
| |||||||
| Optogenetic | Light-sensitive proteins inserted into retinal or cortical neurons | Cell-type specificity; potentially high resolution | Experimental; requires gene therapy and light delivery systems | Nirenberg and Pandarinath (2012) | Potentially full field | Potentially high | Gauvain et al (2021) |
|
| |||||||
| Magnetic, ultrasound | Non-electrical stimulation modalities targeting visual pathways | Possibly less invasive | Early-stage research; unclear spatial/temporal resolution | Lee et al (2016a), Naor et al (2016), Lee et al (2016b) | Unknown | Likely low | None |
Table 2.
Comparison of prosthesis design factors and motion perception. This table summarizes key design features that influence motion perception in visual prosthesis systems. Factors include system resolution, field of view, gaze-contingent processing, motion encoding algorithms, and training or rehabilitation protocols. For each factor, the table outlines its underlying mechanism, potential advantages, and limitations, along with illustrative examples where available. The comparison highlights how engineering trade-offs and system-level choices shape perceptual outcomes in artificial vision.
| Design Factor | Mechanism | Advantages | Limitations | Examples |
|---|---|---|---|---|
| Gaze-contingent systems | Image rendering adjusts in real time based on eye position | Supports spatial updating; improves perceptual stability | Requires eye tracking | Simulated systems (e.g. Paraskvoudi and Pezaris 2021, Rassia et al 2022) |
| Motion encoding algorithms | Artificial enhancement of motion cues | Improves motion detection in prosthetic vision | May introduce unnatural percepts; design-specific trade-offs | Brightness as depth; background subtraction; saliency filtering |
| Training and rehabilitation protocols | Repeated exposure and behavioral training to improve motion perception | Enhances performance over time; low cost | Requires user commitment; real-world performance varies | Orientation and motion training studies |
| Resolution | Limited by electrode count and spacing | Supports basic object recognition | Low resolution; limits reading and faces | Argus II (60); Orion (60); Alpha AMS (1600) |
| Field of view | Set by implant size and layout | Improves spatial awareness | Typically narrow (<20°); trade-offs with resolution | Argus II (20°); PRIMA (7–10°) |
2. Background
Visual motion information originates in the retina and passes through the lateral geniculate nucleus (LGN) before reaching the cortex. The LGN serves as the primary relay station for parvocellular (P), magnocellular (M), and koniocellular (K) processing streams, each of which is specialized for different aspects of visual processing (Solomon 2021). While most motion-related data is found in the LGN’s M layers, P input plays a significant role in shaping spatial, temporal, and chromatic motion cues in various motion-specific brain areas (Nassi et al 2006), and K layers are known to be motion sensitive (Hendry and Reid 2000, Renshaw and Pezaris 2025).
From the LGN, M/P/K pathways project via the optic radiations to the striate cortex (V1) in the occipital lobe. V1, the first stage of integrated visual processing and a focus of visual prosthesis development (Schiller and Tehovnik 2008, Foroushani et al 2018), exhibits retinotopic organization and primarily responds to stimuli within small, well-defined receptive fields (Lian et al 2021). Emerging evidence suggests that V1, as well as other cortical visual centers (V2/3), also respond to motion, particularly interblob cells that are sensitive to object movement (Dupont et al 1994, Seiffert et al 2003).
Visual information progresses anteriorly from V1 to higher-order visual processing centers. A critical center for motion perception is the middle temporal area (V5 or MT, often considered in conjunction with V5A/MST, and stylized as MT+ or the MT complex), located on the posterior edge of the superior temporal sulcus (Born and Bradley 2005). MT+ neurons are highly sensitive to motion direction and speed but unresponsive to object characteristics (Dubner and Zeki 1971, Van Essen et al 1981, Lagae et al 1993). Bilateral MT+ lesions result in impairment of motion detection, speed determination, and direction discrimination (Schiller 1993). Nevertheless, the eventual recovery of some motion perception following bilateral MT+ lesions suggests the involvement of other cortical regions in motion processing despite the primacy of MT+ (Yamasaki and Wurtz 1991).
3. Important distinctions in motion perception through artificial vision
When discussing motion perception through artificial vision, it is important to make two distinctions. The first distinction concerns the difference between movement of phosphenes themselves and the perception of objects that are in motion when viewed through a collection of phosphenes. The second distinction concerns the difference between tasks that specifically assess motion perception, such as drifting grating detection, and tasks which require motion perception for correct completion but do not specifically measure it, such as eye-hand coordination tasks. We will review these two ideas in the following two sections.
3.1. Moving phosphenes versus seeing moving objects through phosphene vision
Phosphene movement is not the same as motion perception derived from phosphene vision. Phosphene movement does not inherently yield motion perception, and motion perception does not necessarily rely on phosphene movement (figure 1). In the visual system, motion and form processing are closely intertwined and can co-occur (Pavan et al 2013), yet one may dominate depending on how visual information is presented (Donato et al 2020).
Figure 1.

Different kinds of motion. Four types of motion are illustrated with before-and-after plots. Each circle represents the perceived location of a phosphene induced by a specific electrode in the prosthesis, relative to the gaze location (red cross); when electrodes are activated, their phosphenes are perceived (filled circles), but when not activated, their phosphenes are not seen (open circles). The top row shows an unlikely scenario where a single illuminated phosphene shifts relative to the rest while the gaze position remains fixed. The second row shows eye movement, causing the entire retinotopically-anchored phosphene field to shift. The third row depicts the gaze being fixed, and thus the phosphene field fixed as well, as a bar moves (dashed lines), changing the subset of illuminated phosphenes representing it, and creating apparent motion. The fourth row depicts the converse condition with gaze moving and the bar fixed. The apparent motion is in the opposite direction, exactly offsetting the gaze shift, so that the brain can interpret the object as stable in the world. Most experiments in this review are variants of the third row. The pattern of phosphenes used in this figure is modeled after measurements from a patient implanted with an Orion device (Oswalt et al 2021). See Paraskevoudi and Pezaris (2018, 2021) for additional detail regarding gaze contingency effects.
Early work by Brindley and Lewin (1968) noted that electrically stimulating the primary visual cortex produced phosphenes which followed eye movements, giving the phosphene the appearance of a fixed position in retinotopic space (that is, always at the same relative position to the point of regard), as confirmed by many later reports (Dobelle and Mladejovsky 1974, Schmidt et al 1996, Caspi et al 2021) for different stages of the early visual pathway (Chapanis et al 1973, Veraart et al 2003, Pezaris and Reid 2007, Caspi et al 2009). Because the visual circuitry compensates for eye position in motion processing, this apparent motion in phosphenes driven by eye position changes is often ineffective at conveying motion within the visual scene unless paired with a system that incorporates eye position in its translation of the scene to phosphene stimulation (Paraskevoudi and Pezaris 2018, 2021).
However, there have also been reports of individual phosphenes that move within the retinotopic field, independent of eye position. Studies show that electrode-generated phosphenes can yield movement and form perception (Wilke et al 2011, Sinclair et al 2016); participants reported perceptions of both form (‘an arc,’ ‘a slightly curved horizontal line’) and motion (‘movement,’ ‘swept across the visual field’) when one or multiple electrodes were activated. These reports seem to be more the exception than the norm, however, and may have been related to confounds such as high levels of ongoing spontaneous phosphenes, and uncontrolled nystagmus. In contrast, form perception was evoked from line tracing of figures in recent research using current-steering to move individual phosphenes (Beauchamp et al 2020). Participants could identify and reproduce patterns created by moving phosphenes, demonstrating how motion alone can convey form information.
The primary means for conveying motion through a phosphene array does not require phosphenes that move, but rather a moving pattern of activation (figure 1). Most current visual prostheses successfully evoke motion perception using electrodes that are fixed in place in the tissue, producing phosphenes that are fixed in retinotopic coordinates (Ayton et al 2014, Chuang et al 2014, Lewis et al 2015). Like with a television, computer screen, or virtual reality headset, successive activation of pixels that themselves do not move can still create the perception of motion, often called apparent motion (Wertheimer 1912, Ramachandran and Anstis 1986). However, technical limitations, such as electrode count, make motion perception challenging in artificial vision (Picaud and Sahel 2014, Stingl et al 2017). For the remainder of this review, we will focus on motion perception evoked by patterned activation and discuss the movement of individual phosphenes only when it intentionally contributes to motion perception.
4. Frontoparallel motion perception in artificial vision
The most basic test of visual motion perception assesses motion detection in the frontoparallel plane (Nakamara and Otsuka 1995). Many studies examining motion perception in artificial vision use tasks similar to the basic assessment of light and motion (BaLM) (Bach et al 2010). The BaLM suite is a validated set of tests that quantify visual acuity in low vision. Of particular interest here is the motion task in which a random pattern of light and dark hexagons begins to move in one of four or eight directions. Subjects indicate the motion’s direction in a four- or eight-alternative forced choice task (figure 2).
Figure 2.
BaLM motion task. (A) The motion detection and identification task from the BaLM suite has patients using their visual prostheses to view moving images on a computer monitor and report the direction seen. Subjects respond with the direction of motion either by verbal identification, by button press, or by touch screen, depending on the study. (Drawing by Zofia Pezaris). (B) Stimuli include motion in one of four directions (first column, in blue) or eight (not shown) and at varying speeds (second column, in green). Image types ordered by complexity include bars, gratings, random dots, and patterns (third column, in orange), although only one image type is typically used in a given experiment. One variation of this test uses a continuous selection of direction, rather than a fixed set of four or eight, with subjects reporting direction via touch screen; the error angle between reported and actual directions is then analyzed. For moving dot stimuli, it should be noted that the moving dots are not themselves phosphenes, but typically an image of a texture that moves rigidly, unlike seminal experiments in MT where the dots may not all move in the same direction (e.g. Britten et al 1992). The motion test in the original BaLM suite used hexagonal patterns like the bottom image in the third column (after figure 1 of Bach et al 2010).
There have been many variations of this task used in the literature, challenging comparability across studies. BaLM adaptations have adopted moving lines or a moving dot pattern rather than the hexagons pattern. Another common variation eliminated the alternative forced choice task, opting instead to have participants indicate the direction of motion on a touchpad (Dorn et al 2013, da Cruz et al 2016) and measuring error as the angular difference between the reported and actual directions of motion. Another variation explored different speeds within the BaLM to quantify the maximum speed at which motion could reliably be detected (Stingl et al 2013b, Stingl et al 2015).
4.1. Frontoparallel motion perception in retinal prostheses
4.1.1. Second sight medical products: Argus I and II Retinal prostheses
The bulk of evidence for motion perception in visual prostheses comes from the Argus retinal prosthesis system, an epiretinal prosthesis developed by second sight medical instruments (Luo and da Cruz 2016). Research comes from the two versions of the implant, the Argus I, with 16 contacts, and the Argus II, with 60 contacts. Both were implanted on the inner retinal surface, designed to stimulate retinal ganglion cells (Humayun et al 1999) with signals from an external camera fixed in a set of goggles (Mirochnik and Pezaris 2019). The Argus II was one of the first visual prosthesis devices to obtain a CE mark (EU regulatory approval) and the only one to obtain FDA (US regulatory) approval, contributing to the preponderance of data with this device (Luo and da Cruz 2016, Farnum and Pelled 2020). Importantly, the camera, being externally worn, does not incorporate eye position into visual field movement to eye position. As naturalistic eye movements are crucial for motion perception in sighted individuals (Dodge 1904, Wallach 1982), eliminating visual field adjustments that track eye position may limit motion perception in both laboratory and daily-use settings (Caspi et al 2018, Paraskevoudi and Pezaris 2021, de Ruyter van Steveninck et al 2024).
Early Argus system research on frontoparallel motion perception had mixed results. A 2003 study using the Argus I (Humayun et al 2003) with one patient showed above chance identification of displacement direction of two sequentially activated phosphenes. The sequential presentation of two nearby phosphenes (within 3 s, average minimum distance about 7° of visual angle; see Humayun et al, figure 2(b)) might not have been expected to trigger apparent motion perception (Julesz and Payne 1968, Anstis 1970, Moleirinho et al 2021), so the perceptual mechanism is unclear. The participant also identified, at above chance levels, the direction of motion from the edge of a 15 cm book entering the image as it was moved up or down 5 cm from the camera; we estimate the book would have spanned the camera image and thus the array’s field of view as well. A second study, also with the Argus I (Yanai et al 2007), found that only one of the three participants could detect movement. The participant’s younger age and shorter duration of visual impairment suggest that other factors, such as progression in retinal degeneration, may impact restoration of motion detection ability. Both studies should be viewed cautiously due to small sample size and lack of methodological validation.
More recent research with Argus II patients used variations of the BaLM assessment, thus increasing methodological standardization and reliability. These studies utilized a direction of motion task in which a high contrast white bar moved across a black touch screen. Performance was quantified using angular response error. Many of these studies found improved motion perception with, compared to without, the device turned on (Humayun et al 2012, Dorn et al 2013, Rizzo et al 2014, Ho et al 2015, da Cruz et al 2016, Arevalo et al 2021). Although statistically significant, the practical significance of the observed improvement should also be considered. While improvement alone suggests the device’s technical capacity to generate phosphenes, evaluating absolute performance may be a better gauge of the device’s functionality as an assistive device.
Accordingly, many of these Argus II studies reported participants’ absolute performance, in addition to significance levels. Even among users experiencing substantial motion perception benefits from the device, absolute error rates remained notably high and variable. Dorn et al (2013) found that participant errors in identifying the direction of motion spanned from around 10° to over 100°, with 18 out of 28 participants falling within the 50°–90° range. Other studies reported mean errors across all participants to range from 55° to 65°, with variations depending on the post-implantation year (Humayun et al 2012, da Cruz et al 2016). Arevalo et al (2021) reported the highest average errors, measuring at 81° with the system on and 91° with it off. In Rizzo et al’s study (2014), the most successful participant accurately identified the travel direction on only 69% trials, however, the definition of a correct response was not specified. The reported error magnitudes, especially when using consistent and validated methodology, cast doubt on these device’s capability of providing motion perception that would be found useful by subjects (Karadima et al 2023). Considering the overarching objective of enhancing the quality of life for visually impaired individuals, motion misperception of these scales raise substantial safety concerns, independent of other limitations, such as narrow field of view, and would likely lead to only marginal improvements in mobility and independence.
Other studies failed to find any significant motion perception benefit using the Argus II device. Studies by Schaffraft et al (2019), Delyfer et al (2021), Yoon et al (2021), and Yue et al (2015), also using variations on the BaLM methodology, found no significant improvement in the ability to identify direction of motion using the Argus II device. When reported, absolute performance for these studies was consistent with absolute performance in studies reporting motion perception benefits. In Schaffraft et al and Delyfer et al, the average response errors were between 60°–70° at each time point (3, 6, 12, 24 months post-implantation).
The discrepancy between studies reporting motion perception benefits with the Argus II and those finding no significant improvement likely arises from differences in experimental design and levels of subject experience. Studies showing benefits (Dorn et al 2013, Rizzo et al 2014, da Cruz et al 2016) often used simplified or repetitive stimuli, extended presentation duration, or restricted directional choices, increasing the likelihood of above-chance performance through compensatory strategy rather than veridical perception. In contrast, studies reporting null results (Yue et al 2015, Schaffrath et al 2019, Delyfer et al 2021, Yoon et al 2021) frequently employed more naturalistic stimuli, shorter exposures, or more rigorous trial randomization, placing greater perceptual demands on subjects. Subject-level factors also likely contributed: individuals who received targeted motion training or reported regular device use were observed to perform better on these tasks (Dorn et al 2013, da Cruz et al 2016), and were hypothesized to have developed compensatory strategies not present in less experienced users (Rizzo et al 2014).
Studies with the Argus II using non-BaLM methodology also find error-prone motion perception performance. In a study by Dagnelie et al (2017), participants indicated the direction of motion of an experimenter walking in front of them (figure 3). While external auditory and tactile cues were reduced, no sensory masking devices were used. With the device engaged, 67% of participants (18/27) indicated the direction of motion with significantly above-chance accuracy. However, of those 18, five also had above-chance performance with the device disengaged. Thus only 48% (13/27) went from non-signficant to significant performance with the device. Even with this 48%, the results should be accepted cautiously given the potential contribution of non-visual, specifically audio, motion information. While there appears to be a strong trend in improvement with the device turned on (25/27 improved), the results were inconsistent across individuals (figure 3(B)).
Figure 3.
Walking motion perception test. One of the non-BaLM tests of motion perception was designed to more closely mimic a real-world experience of detecting people walking. (A) Two testers stood 10 feet apart and 10 feet in front of a subject. At the sound of a tone, one or the other tester walked to the opposite position, and, afterward, the subject reported the perceived direction of motion. (B) Results were mixed; while 25/27 subjects improved on the two-alternative forced choice test (points above the identity diagonal), a minority (13/27, filled blue) improved from chance performance (grey zones) without the device, to statistically above chance with the device (hatched zone); the remainder were already above chance without the device (unfilled red), or had no significant improvement (unfilled gray). (A, image drawn by Zofia Pezaris, after figure 4 of Dagnelie et al 2017; B, data extracted from their figure 9 and replotted).
Overall, the Argus II appears to give rise to increased motion perception. However, this technological success must be carefully contextualized with respect to functional improvements it can offer individuals with visual impairment. The magnitude of motion perception error conferred by Argus II remains too high to support reliable and accurate motion discrimination under realistic visual conditions. Future development and research should prioritize this functional goal. In particular, the integration of eye position tracking with the external camera may be an important feature to add for improving Argus II motion perception, along with the more obvious increases in resolution and field of view.
4.1.2. Retina Implant AG: Alpha IMS and AMS subretinal visual prostheses
Alpha IMS and its successor, Alpha AMS, are two other visual prostheses which have been the focus of research regarding frontoparallel motion detection and assessment. They are subretinal devices developed by Retina Implant AG (Stingl et al 2013b) with integrated imaging and stimulation circuitry placed between the outer layer of the retina and the pigment epithelium to stimulate intermediate layers of the retina, primarily bipolar cells. The benefit of integrating the imaging and stimulation circuitry so that both are fixed to the retina is that an external camera is not needed, as light entering the eye is converted directly to electrical impulses. Thus, users can continue to rely on natural eye movements and physiological retinal processing to steer the field of view, and the device inherently incorporates gaze contingency (Pezaris and Eskandar 2009, Mirochnik and Pezaris 2019). For motion perception, this is important perceptually, as eye movement and motion perception are highly interrelated (e.g. smooth pursuit eye movements for dynamic object tracking, or attention-related saccade eye movements at the onset of motion) (Dodge 1904, Wallach 1982). The Alpha IMS gained CE approval in 2013, while the second-generation Alpha AMS gained CE approval in 2016 (Perriello 2016).
Results from studies on first-generation Alpha IMS’s frontoparallel motion detection are mixed. Most of the studies used variations of the BaLM motion task. However, unlike Argus II studies, which calculated response error in degrees, these studies used two- and four-alternative forced choice designs, making it more difficult to assess and compare absolute performance to the Argus II.
The most promising results come from Stingl et al (2013b), (2013c), (2015), who used a four-alternative forced BaLM motion task with increasing speeds (starting at 3° s−1). They recorded the highest speed at which a participant could identify movement of a random array of dots (chance at 25% correct, passing at 62.5% or better (Stingl et al 2013a). In successive studies with the Alpha IMS, researchers found that 21% (Stingl et al 2015) and 56% (Stingl et al 2013a) of participants were able to accurately identify motion at any speed. Of successful participants, most showed the maximum speed for accurate direction identification between 3° and 7° s−1 (about 5 and 2 s, respectively, to traverse the array diagonally) with one exceptional participant able to accurately identify motion at 35° s−1 (about 0.4 s).
Other studies by Stingl et al provided important data concerning implant location on motion perception, finding that implant placement has significant influence on the patient’s ability to detect motion. A study with 20 participants 2013c) supported foveal placement for the implant for better motion detection, with 75% of participants in the cohort with foveal placement (6/8) able to accurately detect motion at the lowest speed (3° s−1) as compared to the cohort with parafoveal placement who could not detect motion at any speed. The authors in that report suggest that the amount of retinal remodeling, expected to be greater with increasing eccentricity for their patient group, may explain the difference in task ability between the two groups.
The Alpha AMS device improves on its Alpha IMS predecessor through changes in device material aimed at improving prosthesis lifetime, and a slight increase in electrode count from 1500 to 1600 (Daschner et al 2018). Despite these updates, motion detection results using the Alpha AMS and Alpha IMS are roughly equivalent. In 2017, Stingl et al published an interim report on Alpha AMS performance results (2017). Using the same BaLM-based methodology as their previous experiments (Stingl et al 2013b, 2013c, 2015), results showed that motion detection was only possible for two out of 11 patients. In 2018, another study by Edwards et al (2018) used the same motion detection procedures and found that none of the five participants passed the motion detection task despite subfoveal placement. The disparities in motion detection among these studies have not been fully explained but may relate to the level of retinal rewiring as mentioned above.
Like the Argus II, motion perception results from the Alpha IMS and AMS prostheses remain mixed. The introduction of different speeds in the BaLM test revealed a dependence of motion perception on speed of the motion—an important point for better understanding the real-world functional capacity of these devices. However, the use of forced-choice methodology (motion detected or not detected) as opposed to absolute response error (angular difference between stimulus and report) in these studies makes it hard to compare performance between the Argus II and Alpha AMS/IMS devices. Furthermore, performance results as reported in these studies make it hard to draw concrete conclusions about real-world functional performance. Finally, none of these studies measured nor controlled for eye movement, likely giving an advantage to the subretinal Alpha IMS/AMS implants. Future research on eye position and movement during prosthetic vision motion perception will not only help to improve motion perception in subretinal implants but may better inform how non-subretinal devices might use eye position tracking (Caspi et al 2018) to deliver more naturalistic prosthetic vision (Paraskevoudi and Pezaris 2021, de Ruyter van Steveninck et al 2024).
4.1.3. Bionic vision technologies: bionic eye system suprachoroidal visual prosthesis
The bionic eye system is a third retinal prosthesis with published reports of motion detection and perception. The bionic eye system is a suprachoroidal visual prosthesis developed by bionic vision technologies (Ayton et al 2014). The suprachoroidal approach, as opposed to subretinal and epiretinal approaches, has advantages in ease of surgical access (Mirochnik and Pezaris 2019), as the implant is placed between the choroid and the sclera (Villalobos et al 2013), thus taking advantage of the remaining retinal processing that has not been lost to degeneration. Like epiretinal approaches, this device uses an external camera in order to process the visual scene (Mirochnik and Pezaris 2019), but does not include measurement of eye position to compensate phosphene delivery for gaze direction (Pezaris and Eskandar 2009, Paraskevoudi and Pezaris 2021). Currently, this project has not yet received regulatory approval.
Studies with the suprachoroidal implants have been methodologically consistent, relying on the validated BaLM paradigms. This consistency may be attributable to the recency of this research compared to other visual prostheses. The first study on motion perception through a suprachoroidal implant was by Fujikado et al (2011) in 2011, where two participants viewed a white box moving across their visual field either horizontally or vertically. One participant scored significantly higher than chance; the other did not. Another study by Shivdasani et al (2017) in 2017 used similar methods, testing motion perception of a fixed white bar in 8 directions, with 5 randomized speeds. Again, out of the two participants, only one passed the 8-alternative forced choice design, scoring above the 63% passing criterion at 4 out of 5 speeds. The fastest speed of 80 degrees/sec was the only speed at which the more successful participant failed to meet the passing criterion. The second participant was unable to perform at a passing level at any speed, even when the task was simplified to a 4-alternative forced choice paradigm. But they still performed above chance at the slowest speed, accurately resolving motion in 27% or 50% of trials depending on electrical stimulus rate (50 and 400 pulses per second, respectively). With a cohort of four participants, Petoe et al (2021) found similar results: just 2 of 4 participants performed significantly better with the device on (63% and 83% accurate) versus off (28% and 25% accurate) using a 4-alternative forced choice task BaLM task.
Titchener et al (2020) explored the influence of retinotopic map accuracy in motion discrimination when using suprachoroidal devices. Participants were able to perceive the direction of motion of a drifting bar significantly and substantially above chance in normal prosthesis operation (contrast with the drifting grating stimuli used later), with performance deteriorating to near-chance levels with scrambled phosphene location maps, demonstrating the subjects’ use of retinotopic information. Of the three speeds tested (7, 15, 30° s−1), subjects performed better at the lower two. The three subjects exhibited smooth-pursuit-like gaze movements consistent with stimulus direction and speed, as is associated with natural visual perception of motion. However, optokinetic-like eye movements evoked when viewing drifting gratings were highly inconsistent with stimulus motion. The combined findings suggest there was only partial activation of normal motion-processing pathways. The requirement that subjects hold their eyes fixed for normal operation, because of a lack of gaze contingency, complicates interpretation of the results: What was viewed as smooth pursuit eye movements may have been the tracking of phosphenes themselves, rather than the motion of the stimuli.
In summary, the results from the suprachoroidal implant align with those from the Alpha IMS and AMS studies, highlighting the importance of speed in motion perception. However, an interesting observation can be drawn comparing the results from the Alpha IMS/AMS systems to the suprachoroidal devices: Alpha IMS/AMS systems have inherent gaze contingency whereas the suprachoroidal systems do not. Given our understanding of how natural eye movements influence motion perception (as discussed earlier), one might anticipate better performance with the Alpha IMS/AMS devices. However, since gaze position was measured in the suprachoroidal studies but not incorporated into the device operation, and was neither measured nor controlled in the Alpha IMS/AMS studies, interpreting the results in terms of the influence of eye position and movement is challenging. Plausible explanations for the observed similarity is that suprachoroidal participants were instructed to center their gaze and hold it forward, or they may have compensated for the lack of gaze contingency by relying on head position. A study by Titchener et al (2020) identified two participants who exhibited head movements congruent with stimulus motion, suggesting that they may have used external cues, such as head position, to resolve motion.
4.2. Frontoparallel motion perception in cortical prostheses
Cortical prostheses, as opposed to retinal prostheses, stimulate the primary visual cortex to generate phosphenes. Fewer studies have explored motion perception in cortical prostheses than for retinal prostheses. However, the ideas that stimulating the visual cortex could elicit functional phosphenes (Le Roy 1755, Brindley and Lewin 1968, Volta 1997) or that the primary visual cortex is retinotopically mapped (Hubel and Wiesel 1959) are not new. A few important figures in artificial vision have focused their efforts on the development of a cortical prosthesis, including Brindley and Lewin (1968), and Dobelle and Mladejovsky (1974), Dobelle et al 1974, 1976, Dobelle 2000). Since then, other researchers have examined cortical prostheses’ effectiveness in generating motion perception. This review will present two published studies which examine motion detection with cortical devices.
One study by Chen et al (2020) measured motion perception using BaLM-inspired methodology in cortical prostheses. In this study, monkeys were trained to report the direction of motion in sequentially activated dots on a screen. Once the direction of motion task had been learned in natural vision, the stimuli were then presented in phosphene vision. All the monkeys performed significantly above chance in determining direction of motion in phosphene vision, indicating motion perception as well as direction discrimination that was comparable to control conditions with simulated phosphenes viewed through normal vision. It is important to note that the animals maintained central fixation during stimulation in this task, so that eye position was not a confounding factor.
The second motion perception study with a cortical prosthesis was conducted by Beauchamp et al (2020). They studied how activating a sequence of phosphenes, including with current steering, a technique for increasing the placement resolution of individual phosphenes without the potential complications from increased electrode counts (Meikle et al 2022, Meikle and Wong 2022), can elicit form and motion perception (figure 4). Current steering has been used in retinal (Dumm et al 2014, Spencer et al 2018), optic nerve (Yan et al 2016), and cochlear (Firszt et al 2007, Landsberger and Srinivasan 2009) implants, among others. In their experiment, current steering was employed to generate the movement of a perceived phosphene, where the location of a virtual electrode between two physical electrodes was modulated by changing the relative charge at each physical electrode (Meikle et al 2022). The resulting phosphene moved along a specific path, providing form information (i.e. drawing letters) and motion information (i.e. direction of motion). The results showed that the participant performed well above chance, indicating the correct direction of motion 87% of the time (chance was 50%), and recognizing the drawn letter 66% of the time (chance was 25%). Notably, in this study eye position was neither measured nor controlled, although the subjects were instructed to help keep their eyes fixed in space by placing a finger on the monitor in front of them.
Figure 4.

Reproduction of phosphene tracing from cortical stimulation via current steering. Left: retinotopic receptive field maps (blue ovals) for a series of stimulation sites in visual cortex with the order of stimulation marked. Center: schematic depiction of example stimulation sequence across regular (numbers) and virtual (red dots) electrodes. Right: drawing made in response to apparent phosphene motion, showing a trajectory that was recognized by the subject as the letter Z, demonstrating that current steering can convey both form and motion information. A similar test of sequentially activated phosphenes, but often with just two, was used in other studies as well (e.g. Humayun et al 2003, Yanai et al 2007, Chen et al 2020). Reprinted from Beauchamp et al (2020), Copyright (2020), with permission from Elsevier.
The results of motion perception studies in cortical prostheses are promising but limited, with few publications, low sample sizes, and a lack of consistent methodology. To determine the potential for functional motion perception in these devices, more experiments using standardized methods are needed. To maintain consistency and validity, standards such as the BaLM suite developed for use in retinal prostheses (Bach et al 2010) should be applied. Furthermore, as with retinal prostheses, gaze position compensation is an integral requirement to establish motion perception that rises to substantial fractions of normal visual ability. Since phosphenes from cortical stimulation are retinotopically mapped, if eye position is not used by a cortical prosthesis to correct the camera view, phosphenes (always in retinotopic coordinates) will appear to move along with the eyes, confounding motion perception in real-world use.
5. Augmented representation of motion in visual prostheses
We have seen that objectively measuring motion perception in visual prostheses can be challenging, with often unsatisfactory results. One potential solution is to use technology from machine vision, which uses machine learning to extract, analyze, and interpret visual information (Wang et al 2021); with this additional scene analysis, a visual prosthesis might distill camera output in order to enhance the presentation to subjects in specific ways. Although none of the studies reviewed thus far tested the use of machine-vision techniques to specifically enhance motion perception, related filters are already in some visual prosthesis devices, such as color processing, edge detection, and contrast inversion in the Argus II (Gilhooley and Acheson 2017) or image processing algorithms in the IRIS device (not discussed in this review) which mimic retinal temporal resolution (Fauvel and Chalk 2022).
How to best analyze and translate raw scene data for machine-vision enhanced visual prostheses, the encoding problem, remains an open question (reviewed in Wang et al 2021). Of the attempts to use machine vision methods specifically for motion perception in visual prostheses, the majority of studies have focused on navigation tasks. As other papers have done comprehensive reviews on machine vision in visual prostheses (Barnes 2012, Perez-Yus et al 2017, Wang et al 2021, de Ruyter van Steveninck et al 2022), we will concentrate on motion perception alone.
Two common scene processing strategies from machine vision with application to visual prostheses are intensity filtering and depth filtering. Intensity-based visual presentation, or veridical encoding, is how information is classically presented in visual prostheses; objects which emit more light to the camera are represented through brighter phosphenes. Depth filtering uses brightness to encode closeness to the camera; objects that are closer in the scene appear brighter in phosphene vision. Thus, objects moving toward a user, which are a potential threat, become brighter as well as larger, whereas objects moving away from a user become dimmer as well as smaller. Although depth filtering fundamentally enhances position, rather than motion directly, the change in position (and representative change using depth filtering) may be useful in detecting looming motion, for example.
Research examining the impact of depth filtering on motion perception in visual prostheses has shown that depth filtering is effective in obstacle avoidance tasks, especially in environments with hanging obstacles, and outperforms intensity filtering in some cases (Barnes et al 2011, Lieby et al 2011). In a 2020 study (Kartha et al 2020) employing the Argus II, the incorporation of depth filtering markedly enhanced motion direction detection in a task assessing the direction of motion of individuals on an escalator from 5–6 m away. Detection rates were 6%–21% (depending on the participant) without depth filtering and increased to between 69% and 73% with depth filtering. Other studies have used more advanced depth filtering algorithms, such as visually differentiating between the ground and other non-ground surfaces, which further improves obstacle avoidance performance (McCarthy et al 2014).
Another possible technique is background subtraction. Background subtraction is similar to depth filtering. In this paradigm, rather than depth being presented as a gradient of brightness, objects in the foreground are isolated from the background and presented alone. One study of sighted subjects viewing a simulation of artificial vision found that background subtraction of dynamic stimuli in a motion detection task significantly improved performance from about 50% to above 75% correct (Wang et al 2014). This type of filtering is useful for perception of individual moving objects (such as frontoparallel motion) or other cases where camera location remains fixed, however, it may impede performance in navigation or other self-motion paradigms.
A more complicated image filtering technique called saliency filtering, uses a combination of intensity, depth, saturation, and edge detection to brighten the most salient or prominent visual images. Combined, these factors highlight important information in the scene better than depth filtering alone can. Studies examining saliency filtering have found that these representations appear subjectively more useful to sighted viewers (Mccarthy et al 2013) and can have marked benefit to obstacle avoidance in new environments (Parikh et al 2013).
Other machine vision techniques aimed at improving motion perception in visual prostheses are free space filtering and time-to-contact representation. Free space filtering highlights free, traversable space for the user based on information about their movement, orientation, and objects in the scene, and is well-suited for navigation tasks (Perez-Yus et al 2017). Time-to-contact representation, on the other hand, uses a ratio of relative velocity and distance to determine object brightness, allowing for better threat detection by making faster or closer moving objects brighter (McCarthy and Barnes 2012). While this method has broader motion perception benefits, it may not effectively represent frontoparallel motion, as objects moving orthogonal to the observer are not more noticeable than stationary objects at the same distance. These two techniques have not been tested in clinical settings.
Thermal imaging, when incorporated into visual prostheses, has shown considerable promise for enhancing motion perception (Sadeghi et al 2021) and supporting mobility (He et al 2020, Sadeghi et al 2024) in artificial vision. Specifically, thermal sensors provide a signal that highlights warm objects—such as humans, animals, or vehicles—even in low-light or visually complex environments, thereby improving salience and aiding in object tracking. For instance, Sadeghi et al (2021) demonstrated that integrating thermal imagery allowed prosthesis users to better detect and follow people (warm objects) walking in building interiors (cooler backgrounds). These findings suggest that thermal imaging and perhaps other spectral enhancements may be valuable additions to artificial vision systems and should be further explored for their potential to improve functional outcomes in real-world settings.
6. Discussion
The results surveyed in this review underscore a striking variability in motion perception performance across individuals and task designs, which has important implications for future visual prosthesis development. While some participants exhibit marked capabilities with certain devices or stimulus conditions, others show little or no perceptual benefit, even under similar experimental protocols. This inconsistency highlights shortcomings in our knowledge of how device characteristics, placement variability, and residual visual pathway integrity interact to provide interpretable input to the dorsal stream and, critically, what characteristics of that input make it interpretable. We encourage continued adoption of the BaLM suite to standardize motion perception assessments while also recognizing that an expansion of that suite to functionally relevant benchmarks that incorporate wider contextual frameworks is necessary.
The variability of results also echoes the complexity inherent in replicating natural motion perception through artificial vision. In the discussion that follows, we explore the implications of these findings, emphasizing the importance of standardized and contextually meaningful assessment measures, the integration of gaze position tracking, and emerging technological developments that hold promise for substantially enhancing prosthetic vision. Additionally, we propose experimental directions aimed at deepening our understanding of the neural mechanisms underlying artificial motion perception and improving functional outcomes for prosthesis users.
6.1. Measurements and reporting
There are two distinct measures by which one can assess quantitative success in motion perception for visual prostheses in laboratory settings: statistical significance and absolute performance. While each of these measures have strengths and weaknesses, they are fundamentally different. Each of these two measures will be considered here, bearing in mind that a well-designed experiment will create a homology between statistical significance and a generalizable real-world outcome.
6.1.1. Statistical significance
Clinical studies have specific criteria for success that often include statistically significant improvement in some measurement. But from a scientific standpoint, it is important to consider whether measured differences are substantial as well as significant: do the results convey or demonstrate an understanding of a phenomenon or mechanism? For example, individuals with retinitis pigmentosa (RP, a common blinding disease) can have very poor motion perception compared to normally-sighted individuals and profoundly blind individuals have none (Turano and Wang 1992). A hypothetical improvement in motion perception from blindness to the level of RP function can be significant, without being useful. It is important to examine both the significance of measured differences as well as their magnitude in comparison to normal vision to understand the potential implications for device utility, as prospective users concentrate on functional ability, such as recognizing individuals or detecting objects for safe mobility, not statistical significance (Keeffe et al 2010, Lane et al 2011, 2012, Karadima et al 2023)
6.1.2. Absolute performance and contextualization
Across the studies presented here, there was limited agreement on absolute performance. Performance ranged from 6% to 100% based on the specific task or individual, with only a few studies providing explicit thresholds for success. Some studies utilized chance level or forced choice tasks as a method of determining success. Although the forced choice paradigm is a well-validated tool for assessing absolute performance and provides a specific success threshold (Zavala 1965, Burgess 1995, Jäkel and Wichmann 2006, Stingl et al 2013a), those success thresholds should be contextualized by their functional relevance. For example, visually assessing the speed of a baseball with 75% accuracy might translate to a successful career in sports media, while visually assessing the relative speed of an incoming car with the same 75% accuracy might result in a fatal outcome. Accordingly, examining raw absolute performance data alone in the absence of any contextualization makes it hard to describe to prospective users what tasks they will and will not be able to do with a given device. Nuance and detail is necessary as long as artificial vision remains less capable than normal vision.
6.2. Gaze contingency
The integration of gaze position in visual prosthesis either inherently in the design, or by addition of gaze tracking, will improve motion perception in visual prosthesis (Caspi et al 2018, Paraskevoudi and Pezaris 2018, 2021, de Ruyter van Steveninck et al 2024). Many aspects of motion perception, including tracking moving objects as well as creating a spatial representation of movement (Thomas and Lleras 2007), are linked to eye movements. Therefore, visual prostheses which can incorporate natural eye movements have a greater chance of providing useful motion perception to the user.
Visual prosthesis systems differ markedly in whether and how they account for eye movements (figure 5). In the first class of devices, including many subretinal implants, the imaging and stimulating surfaces are combined and fixed to the retina and thus move naturally with the eye so they automatically compensate for eye movements (figure 5(A)). These devices require no patient training on how to control gaze, as the prosthetic interface closely replicates the function of eye movements. In the second class, systems use an external scene camera that lacks any means of adjusting the camera field of view to track the gaze position (figure 5(B)), breaking the alignment between visual percepts and world coordinates. Devices without gaze compensation require the patient to suppress eye movements, fixing their eyes in an unnatural, straight-ahead position, and scan the scene with head movements alone, with significant, negative impacts on user experience and outcomes (Erickson-Davis and Korzybska 2021). In the third class, an eye-tracking camera is incorporated to shift a region of interest (ROI) from the camera image, so that phosphene activity matches the gaze view, providing a stable presentation of the external world (figure 5(C)). Although not yet widely implemented, this architecture can be used with retinal, optic-nerve, thalamic, or cortical stimulating arrays to create gaze-contingent artificial vision. Each configuration fundamentally shapes perceptual experience, particularly for motion processing, and help clarify why seemingly similar devices can lead to divergent functional outcomes.
Figure 5.
Functional configurations of visual prosthesis systems with respect to gaze compensation. (A) Subretinal prostheses with integrated photodiodes and stimulating electrodes preserve gaze contingency naturally: the retinal image moves with the eye, maintaining world-referenced stability of phosphenes once the brain applies the efference copy of eye position to retinal output through the visual pathway. (B) Epiretinal prostheses using a fixed scene camera lack gaze information, so the retinal image remains fixed while the eye moves, causing misalignment between gaze and phosphene position. (C) Systems that incorporate an eye-tracking camera and active compensation can restore alignment by selecting a region of interest (ROI) from the camera output based on gaze direction and modifying stimulation delivered to the electrode array to be centered on that area. Any approach enhanced in this way will provide signals to the brain that more accurately reflect the visual changes normally produced by voluntary movement of the eyes, thereby improving device usability.
However, even with more phosphenes, better machine vision algorithms, or more refined biotechnological approaches for implantation and stimulation, the study of motion perception in visual prostheses will remain impeded by an inability to effectively assess and compare performance. This diminishment directly results from a lack of methodological consistency which is especially apparent in current real-world motion perception tasks. Tasks should be methodologically valid and contextualized so that performance can be compared across tasks and devices and, importantly, understood in comparison to normal vision. A given hypothetical device will likely provide a wide range of absolute performance depending on the specific task (e.g. a direction-discrimination versus a motion-speed-threshold task). This variation may be attributable to the amount or type of motion perception required for task performance, or the specific presentation of that motion. Thus, a battery of tasks may prove the most useful, not only to provide fundamental insight into the perception of motion through artificial vision and the mechanisms of engagement of higher-order visual areas, but also to assess a specific device’s utility across several different scenarios. It remains paramount that individual methodologies be rigorously motivated, validated, quantified, and, where possible, standardized.
6.3. Emerging technologies
Near-future advances in visual prosthesis development hold substantial promise for improving motion perception by addressing current technical limitations. One example is PRIMA (Palanker et al 2020), a very high resolution subretinal device implanted in perifoveal areas of geographic atrophy in macular degeneration; with integrated imaging and stimulation circuitry, it naturally incorporates eye movements. While outcomes have not yet been reported, it has the potential for excellent motion perception through a design that fuses artificial vision with remaining peripheral natural vision (Palanker et al 2022). Similarly, the Phosphoenix thalamic prosthesis, currently in pre-clinical development, targets the LGN with greater precision and resolution than the devices discussed above. Across all devices, increasing electrode density and improving spatiotemporal encoding are key to refining motion cues, especially for fast or complex trajectories. Coupled with improvements in gaze tracking and advanced vision algorithms, these innovations may enable prosthetic systems to support functionally meaningful motion perception. Still, assuming that better technical capacity alone will provide better performance may be misguided; the results we have reviewed above suggest that fundamental gaps remain in our understanding of the basic science of perception through artificial vision.
6.4. Proposed experiments
Based on the findings of this review, we propose two lines of inquiry to further the understanding of motion perception through visual prostheses. The first is to develop our knowledge of how artificial vision activates the dorsal stream. We propose exploiting the motion aftereffect, a perceptual illusion in which prolonged viewing of motion in one direction causes a subsequent stationary pattern to appear to drift in the opposite direction, reflecting adaptation in motion-sensitive neurons of the MT+ complex (Tootell et al 1995, He et al 1998, Huk et al 2001). By examining the depth of effect, we can assess the perceptual strength of motion relative to normal vision. This task (figure 6) is amenable to presentation in implanted devices, simulations of artificial vision with normally-sighted subjects as depicted in the figure, and pre-clinical animal experiments. The second line of inquiry is to examine non-frontoparallel motion by investigating looming perception, the ability to detect and interpret approaching objects as their visual size increases, also expected to engage MT+. By examining perceptual thresholds for looming (figure 7), with parameters that mimic what a pedestrian might expect to see from motor vehicles while walking in an urban environment, we again can compare performance to normal vision, including phenomena like accumulation of information as successive motion cues are integrated, while also creating a baseline expectation for perceptual limits of specific devices.
Figure 6.
Motion aftereffect task. This task is designed to elicit a motion aftereffect (MAE) by presenting a prolonged adaptating stimulus followed by a perceptual probe. (A) Participants view a drifting grating (here shown through phosphene vision) for an extended period, holding their gaze fixed. This first adapting stimulus habituates motion circuitry in the MT/MST complex that, like afterimages, can influence the perception of subsequent stimuli. (B) After the adapting stimulus, participants briefly view a probe stimulus (here shown in normal vision) and (C) report its apparent motion, for example through a keypress. Probe stimuli typically drift at lower speeds than adapting stimuli (indicated by shorter arrows), and could also be static. Any shift of the psychometric curve fitted to probe responses versus the actual probe speed and direction reflects the depth of habituation within the MT/MST complex, and thus the level of perception of motion afforded by artificial vision.
Figure 7.
Looming motion discrimination task. This task is designed to measure the ability to discern motion of an object toward or away from the viewer. (A) Participants are shown a static object (here in normal vision. (B) After a variable amount of time, the object begins to change size, simulating an object that has started looming toward or away from the observer. (C) Participants indicate when the object appeared to approach or recede via a button press. The task probes motion-in-depth perception, and can be used under conditions designed to emulate dynamic changes in distance or size during motion-in-depth.
7. Conclusion
Restoring motion perception through visual prostheses remains a significant challenge, with current devices demonstrating limited and variable success across individuals. While statistically significant improvements have been observed in many studies, absolute performance often falls short of levels required for functional tasks like navigation or object tracking. These discrepancies reflect both technological limitations and gaps in our understanding of how artificial vision engages the neural substrates of motion perception. In particular, the extent to which prosthetic stimulation activates the dorsal visual stream, beginning with area MT+, remains largely unexplored. As MT+ is central to motion integration and discrimination, clarifying how artificial stimulation recruits this area, and subsequent downstream areas, is critical to advancing the field.
A coherent picture of motion perception in prosthetic vision will require standardized methodologies, improved measurement of eye position, and task designs that probe higher-order processing. For engineers and designers, incorporating gaze contingency and scene-aware image filtering (e.g. depth, saliency, and thermal cues) may better approximate naturalistic vision. Researchers should prioritize experiments that quantify engagement of specific dorsal stream structures, including the suggested motion aftereffect and looming detection tasks. Clinicians must be prepared to communicate the distinction between statistical improvement and real-world functionality, while helping patients set expectations aligned with current device capabilities. Despite the current limitations, the trajectory of innovation across neural engineering, machine vision, and neuroscience suggests a future in which motion perception can be meaningfully restored above and beyond mere improvements in phosphene count. Achieving that goal depends on deepening our understanding of how artificial vision engages the brain’s motion-processing systems.
Acknowledgments
This research was supported by: National Eye Institute under EY027888; William M. Wood Foundation, Bank of America, Trustee; NIMA Foundation; Peter and Yayi Pezaris; and Don Good.
Data availability statement
No new data were created or analysed in this study.
References
- Anstis S M. Phi movement as a subtraction process. Vis. Res. 1970;10:1411–IN5. doi: 10.1016/0042-6989(70)90092-1. [DOI] [PubMed] [Google Scholar]
- Arevalo J F, et al. for the KKESH Collaborative Retina Study Group Argus II retinal prosthesis for retinitis pigmentosa in the middle East: the 2015 Pan-American association of ophthalmology gradle lecture. Int. J. Retina Vitr. 2021;7:65. doi: 10.1186/s40942-021-00324-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ayton L N, et al. for the Bionic Vision Australia Research Consortium First-in-human trial of a novel suprachoroidal retinal prosthesis. PLoS One. 2014;9:e115239. doi: 10.1371/journal.pone.0115239. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bach M, Wilke M, Wilhelm B, Zrenner E, Wilke R. Basic quantitative assessment of visual performance in patients with very low vision. Invest. Ophthalmol. Vis. Sci. 2010;51:1255–60. doi: 10.1167/iovs.09-3512. [DOI] [PubMed] [Google Scholar]
- Barnes N. The role of computer vision in prosthetic vision. Image Vis. Comput. 2012;30:478–9. doi: 10.1016/j.imavis.2012.05.007. [DOI] [Google Scholar]
- Barnes N, Lieby P, Dennet H, Walker J, McCarthy C, Liu N, Li Y. Investigating the role of single-viewpoint depth data in visually-guided mobility. J Vis. 2011;11:926. doi: 10.1167/11.11.926. [DOI] [Google Scholar]
- Beauchamp M S, Oswalt D, Sun P, Foster B L, Magnotti J F, Niketeghad S, Pouratian N, Bosking W H, Yoshor D. Dynamic stimulation of visual cortex produces form vision in sighted and blind humans. Cell. 2020;181:774–83.e5. doi: 10.1016/j.cell.2020.04.033. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Born R T, Bradley D C. Structure and function of visual area MT. Annu. Rev. Neurosci. 2005;28:157–89. doi: 10.1146/annurev.neuro.26.041002.131052. [DOI] [PubMed] [Google Scholar]
- Bourne R R A, et al. Global prevalence of blindness and distance and near vision impairment in 2020: progress towards the vision 2020 targets and what the future holds. Invest. Ophthalmol. Vis. Sci. 2020;61:2317. [Google Scholar]
- Brindley G S, Lewin W S. The sensations produced by electrical stimulation of the visual cortex. J. Physiol. 1968;196:479–93. doi: 10.1113/jphysiol.1968.sp008519. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Britten K H, Shadlen M N, Newsome W T, Movshon J A. The analysis of visual motion: a comparison of neuronal and psychophysical performance. J. Neurosci. 1992;12:4745–65. doi: 10.1523/JNEUROSCI.12-12-04745.1992. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Burgess A E. Comparison of receiver operating characteristic and forced choice observer performance measurement methods. Med. Phys. 1995;22:643–55. doi: 10.1118/1.597576. [DOI] [PubMed] [Google Scholar]
- Caspi A, Barry M P, Patel U K, Salas M A, Dorn J D, Roy A, Niketeghad S, Greenberg R J, Pouratian N. Eye movements and the perceived location of phosphenes generated by intracranial primary visual cortex stimulation in the blind. Brain Stimul. 2021;14:851–60. doi: 10.1016/j.brs.2021.04.019. [DOI] [PubMed] [Google Scholar]
- Caspi A, Dorn J D, McClure K H, Humayun M S, Greenberg R J, McMahon M J. Feasibility study of a retinal prosthesis: spatial vision with a 16-electrode implant. Arch. Ophthalmol. 2009;127:398–401. doi: 10.1001/archophthalmol.2009.20. [DOI] [PubMed] [Google Scholar]
- Caspi A, Roy A, Wuyyuru V, Rosendall P E, Harper J W, Katyal K D, Barry M P, Dagnelie G, Greenberg R J. Eye movement control in the Argus II retinal-prosthesis enables reduced head movement and better localization precision. Invest. Ophthalmol. Vis. Sci. 2018;59:792–802. doi: 10.1167/iovs.17-22377. [DOI] [PubMed] [Google Scholar]
- Chapanis N P, Uematsu S, Konigsmark B, Walker A E. Central phosphenes in man: a report of three cases. Neuropsychologia. 1973;11:1–19. doi: 10.1016/0028-3932(73)90059-6. [DOI] [PubMed] [Google Scholar]
- Chen X, Wang F, Fernandez E, Roelfsema P R. Shape perception via a high-channel-count neuroprosthesis in monkey visual cortex. Science. 2020;370:1191–6. doi: 10.1126/science.abd7435. [DOI] [PubMed] [Google Scholar]
- Chuang A T, Margo C E, Greenberg P B. Retinal implants: a systematic review. Br. J. Ophthalmol. 2014;98:852–6. doi: 10.1136/bjophthalmol-2013-303708. [DOI] [PubMed] [Google Scholar]
- da Cruz L, et al. Five-year safety and performance results from the Argus II retinal prosthesis system clinical trial. Ophthalmology. 2016;123:2248–54. doi: 10.1016/j.ophtha.2016.06.049. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dagnelie G, et al. the Argus® II Study Group Performance of real-world functional vision tasks by blind subjects improves after implantation with the Argus® II retinal prosthesis system. Clin. Exp. Ophthalmol. 2017;45:152–9. doi: 10.1111/ceo.12812. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Daschner R, Rothermel A, Rudorf R, Rudorf S, Stett A. Functionality and performance of the subretinal implant chip Alpha AMS. Sens. Mater. 2018;179:179–92. doi: 10.18494/SAM.2018.1726. [DOI] [Google Scholar]
- Delyfer M-N, Gaucher D, Mohand-Saïd S, Barale P-O, Rezaigua-Studer F, Ayello-Scheer S, Dollfus H, Dorn J D, Korobelnik J-F, Sahel J-A. Improved performance and safety from Argus II retinal prosthesis post-approval study in France. Acta Ophthalmol. 2021;99:e1212–e21. doi: 10.1111/aos.14728. [DOI] [PubMed] [Google Scholar]
- Dobelle W H. Artificial vision for the blind by connecting a television camera to the visual cortex. ASAIO J. 2000;46:3–9. doi: 10.1097/00002480-200001000-00002. [DOI] [PubMed] [Google Scholar]
- Dobelle W H, Mladejovsky M G. Phosphenes produced by electrical stimulation of human occipital cortex, and their application to the development of a prosthesis for the blind. J. Physiol. 1974;243:553–76. doi: 10.1113/jphysiol.1974.sp010766. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dobelle W H, Mladejovsky M G, Evans J R, Roberts T S, Girvin J P. ‘Braille’ reading by a blind volunteer by visual cortex stimulation. Nature. 1976;259:111–2. doi: 10.1038/259111a0. [DOI] [PubMed] [Google Scholar]
- Dobelle W H, Mladejovsky M G, Girvin J P. Artifical vision for the blind: electrical stimulation of visual cortex offers hope for a functional prosthesis. Science. 1974;183:440–4. doi: 10.1126/science.183.4123.440. [DOI] [PubMed] [Google Scholar]
- Dodge R. The participation of the eye movements in the visual perception of motion. Psychol. Rev. 1904;11:1–14. doi: 10.1037/h0071641. [DOI] [Google Scholar]
- Donato R, Pavan A, Campana G. Investigating the interaction between form and motion processing: a review of basic research and clinical evidence. Front. Psychol. 2020;11:566848. doi: 10.3389/fpsyg.2020.566848. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dorn J D, Ahuja A K, Caspi A, da Cruz L, Dagnelie G, Sahel J-A, Greenberg R J, McMahon M J, for the Argus II Study Group The detection of motion by blind subjects with the epiretinal 60-electrode (Argus II) retinal prosthesis. JAMA Ophthalmol. 2013;131:183–9. doi: 10.1001/2013.jamaophthalmol.221. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dubner R, Zeki S M. Response properites and receptive fields of cells in an anatomically defined region of the superior temporal sulcus in the monkey. Brain Res. 1971;35:528–32. doi: 10.1016/0006-8993(71)90494-X. [DOI] [PubMed] [Google Scholar]
- Dumm G, Fallon J B, Williams C E, Shivdasani M N. Virtual electrodes by current steering in retinal prostheses. Invest. Ophthalmol. Vis. Sci. 2014;55:8077–85. doi: 10.1167/iovs.14-15391. [DOI] [PubMed] [Google Scholar]
- Dupont P, Orban G A, De Bruyn B, Verbruggen A, Mortelmans L. Many areas in the human brain respond to visual motion. J. Neurophysiol. 1994;72:1420–4. doi: 10.1152/jn.1994.72.3.1420. [DOI] [PubMed] [Google Scholar]
- Eckert K A, Lansingh V C, Carter M J, Frick K D. Update of a simple model to calculate the annual global productivity loss due to blindness and moderate and severe vision impairment. Ophthalmic Epidemiol. 2022;30:1–9. doi: 10.1080/09286586.2022.2072899. [DOI] [PubMed] [Google Scholar]
- Edwards T L, Cottriall C L, Xue K, Simunovic M P, Ramsden J D, Zrenner E, MacLaren R E. Assessment of the electronic retinal implant Alpha AMS in restoring vision to blind patients with end-stage retinitis pigmentosa. Ophthalmology. 2018;125:432–43. doi: 10.1016/j.ophtha.2017.09.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Erickson-Davis C, Korzybska H. What do blind people “see” with retinal prostheses? Observations and qualitative reports of epiretinal implant users. PLoS One. 2021;16:e0229189. doi: 10.1371/journal.pone.0229189. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Evans J R, Fletcher A E, Wormald R P L. Depression and anxiety in visually impaired older people. Ophthalmology. 2007;114:283–8. doi: 10.1016/j.ophtha.2006.10.006. [DOI] [PubMed] [Google Scholar]
- Farnum A, Pelled G. New vision for visual prostheses. Front. Neurosci. 2020;14 doi: 10.3389/fnins.2020.00036. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fauvel T, Chalk M. Human-in-the-loop optimization of visual prosthetic stimulation. J. Neural Eng. 2022;19:036038. doi: 10.1088/1741-2552/ac7615. [DOI] [PubMed] [Google Scholar]
- Firszt J B, Koch D B, Downing M, Litvak L. Current steering creates additional pitch percepts in adult cochlear implant recipients. Otol. Neurotol. 2007;28:629. doi: 10.1097/01.mao.0000281803.36574.bc. [DOI] [PubMed] [Google Scholar]
- Foroushani A N, Pack C C, Sawan M. Cortical visual prostheses: from microstimulation to functional percept. J. Neural Eng. 2018;15:021005. doi: 10.1088/1741-2552/aaa904. [DOI] [PubMed] [Google Scholar]
- Fujikado T, et al. Testing of semichronically implanted retinal prosthesis by suprachoroidal-transretinal stimulation in patients with retinitis pigmentosa. Invest. Ophthalmol. Vis. Sci. 2011;52:4726–33. doi: 10.1167/iovs.10-6836. [DOI] [PubMed] [Google Scholar]
- Gauvain G, et al. Optogenetic therapy: high spatiotemporal resolution and pattern discrimination compatible with vision restoration in non-human primates. Commun. Biol. 2021;4:125. doi: 10.1038/s42003-020-01594-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gilhooley M J, Acheson J. Artificial vision: principles and prospects. Curr. Opin. Neurol. 2017;30:55–60. doi: 10.1097/WCO.0000000000000412. [DOI] [PubMed] [Google Scholar]
- He S, Cohen E R, Hu X. Close correlation between activity in brain area MT/V5 and the perception of a visual motion aftereffect. Curr. Biol. 1998;8:1215–8. doi: 10.1016/S0960-9822(07)00512-X. [DOI] [PubMed] [Google Scholar]
- He Y, Sun S Y, Roy A, Caspi A, Montezuma S R. Improved mobility performance with an artificial vision therapy system using a thermal sensor. J. Neural Eng. 2020;17:045011. doi: 10.1088/1741-2552/aba4fb. [DOI] [PubMed] [Google Scholar]
- Hendry S H C, Reid R C. The koniocellular pathway in primate vision. Annu. Rev. Neurosci. 2000;23:127–53. doi: 10.1146/annurev.neuro.23.1.127. [DOI] [PubMed] [Google Scholar]
- Ho A C, et al. Long-term results from an epiretinal prosthesis to restore sight to the blind. Ophthalmology. 2015;122:1547–54. doi: 10.1016/j.ophtha.2015.04.032. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hubel D H, Wiesel T N. Receptive fields of single neurones in the cat’s striate cortex. J. Physiol. 1959;148:574–91. doi: 10.1113/jphysiol.1959.sp006308. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Huk A C, Ress D, Heeger D J. Neuronal basis of the motion aftereffect reconsidered. Neuron. 2001;32:161–72. doi: 10.1016/s0896-6273(01)00452-4. [DOI] [PubMed] [Google Scholar]
- Humayun M S, et al. Visual perception in a blind subject with a chronic microelectronic retinal prosthesis. Vis. Res. 2003;43:2573–81. doi: 10.1016/S0042-6989(03)00457-7. [DOI] [PubMed] [Google Scholar]
- Humayun M S, de Juan Jr E, Weiland J D, Dagnelie G, Katona S, Greenberg R, Suzuki S. Pattern electrical stimulation of the human retina. Vis. Res. 1999;39:2569–76. doi: 10.1016/s0042-6989(99)00052-8. [DOI] [PubMed] [Google Scholar]
- Humayun M S, et al. for the Argus II Study Group Interim results from the international trial of Second Sight’s visual prosthesis. Ophthalmology. 2012;119:779–88. doi: 10.1016/j.ophtha.2011.09.028. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jäkel F, Wichmann F A. Spatial four-alternative forced-choice method is the preferred psychophysical method for naïve observers. J. Vis. 2006;6:13. doi: 10.1167/6.11.13. [DOI] [PubMed] [Google Scholar]
- Julesz B, Payne R A. Differences between monocular and binocular stroboscopic movement perception. Vis. Res. 1968;8:433–44. doi: 10.1016/0042-6989(68)90111-9. [DOI] [PubMed] [Google Scholar]
- Karadima V, Pezaris E A, Pezaris J S. Attitudes of potential recipients toward emerging visual prosthesis technologies. Sci. Rep. 2023;13:10963. doi: 10.1038/s41598-023-36913-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kartha A, Sadeghi R, Barry M P, Bradley C, Gibson P, Caspi A, Roy A, Dagnelie G. Prosthetic visual performance using a disparity-based distance-filtering system. Transl. Vis. Sci. Technol. 2020;9:27. doi: 10.1167/tvst.9.12.27. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Keeffe J E, Francis K L, Luu C D, Barnes N, Lamoureux E L, Guymer R H. Expectations of a visual prosthesis: perspectives from people with impaired vision. Invest. Ophthalmol. Vis. Sci. 2010;51:6020. [Google Scholar]
- Köberlein J, Beifus K, Schaffert C, Finger R P. The economic burden of visual impairment and blindness: a systematic review. BMJ Open. 2013;3:e003471. doi: 10.1136/bmjopen-2013-003471. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Krumpaszky H G, Klauss V. Epidemiology of blindness and eye disease. Ophthalmologica. 1996;210:1–84. doi: 10.1159/000310663. [DOI] [PubMed] [Google Scholar]
- Lagae L, Raiguel S, Orban G A. Speed and direction selectivity of macaque middle temporal neurons. J. Neurophysiol. 1993;69:19–39. doi: 10.1152/jn.1993.69.1.19. [DOI] [PubMed] [Google Scholar]
- Landsberger D M, Srinivasan A G. Virtual channel discrimination is improved by current focusing in cochlear implant recipients. Hear. Res. 2009;254:34–41. doi: 10.1016/j.heares.2009.04.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lane F J, Huyck M H, Troyk P. Looking ahead: planning for the first human intracortical visual prosthesis by using pilot data from focus groups of potential users. Disabil. Rehabil. 2011;6:139–47. doi: 10.3109/17483107.2010.514381. [DOI] [PubMed] [Google Scholar]
- Lane F J, Huyck M, Troyk P, Schug K. Responses of potential users to the intracortical visual prosthesis: final themes from the analysis of focus group data. Disabil. Rehabil. 2012;7:304–13. doi: 10.3109/17483107.2011.629326. [DOI] [PubMed] [Google Scholar]
- Le Roy C. Ou l’on rend compte de quelques tentatives que l’on a faites pour guerir plusieurs maladies par l’electricite. Hist. Acad. Roy. Sci. Mém. Math. Phys. (Mémoires) 1755;60:60–98. (Source: gallica.bnf.fr / Bibliothèque Nationale de France) [Google Scholar]
- Lee S Woo, Fallegger F, Casse B D, Fried S I. Implantable microcoils for intracortical magnetic stimulation. Sci. Adv. 2016a;2:e1600889. doi: 10.1126/sciadv.1600889. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lee W, Kim H, Jung Y, Chung Y An, Song I, Lee J, Yoo S. Transcranial focused ultrasound stimulation of human primary visual cortex. Sci. Rep. 2016b;6:34026. doi: 10.1038/srep34026. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lewis P M, Ackland H M, Lowery A J, Rosenfeld J V. Restoration of vision in blind individuals using bionic devices: a review with a focus on cortical visual prostheses. Brain Res. 2015;1595:51–73. doi: 10.1016/j.brainres.2014.11.020. [DOI] [PubMed] [Google Scholar]
- Lian Y, Almasi A, Grayden D B, Kameneva T, Burkitt A N, Meffin H, Einhäuser W. Learning receptive field properties of complex cells in V1. PLoS Comput. Biol. 2021;17:e1007957. doi: 10.1371/journal.pcbi.1007957. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lieby P, Barnes N, McCarthy C, Liu N, Dennett H, Walker J G, Botea V, Scott A F. Substituting depth for intensity and real-time phosphene rendering: visual navigation under low vision conditions. Annual Int. Conf. IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual Int. Conf. 2011; 2011. pp. 8017–20. [DOI] [PubMed] [Google Scholar]
- Lim Z W, et al. Association between visual impairment and decline in cognitive function in a multiethnic asian population. JAMA Network Open. 2020;3:e203560. doi: 10.1001/jamanetworkopen.2020.3560. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Luo Y H-L, da Cruz L. The Argus® II retinal prosthesis system. Prog. Retin. Eye Res. 2016;50:89–107. doi: 10.1016/j.preteyeres.2015.09.003. [DOI] [PubMed] [Google Scholar]
- Makawi A, Shah V, Hao Tang P, Kim L A, Akkara J D, Montezuma S. Retina prosthesis EyeWiki. 2024. (available at: https://eyewiki.org/Retina_Prosthesis#Photovoltaic_Retinal_Implant_.28PRIMA.29_bionic_vision_system)
- McCarthy C, Barnes N. Time-to-contact maps for navigation with a low resolution visual prosthesis. 2012 Annual Int. Conf. IEEE Engineering in Medicine and Biology Society; 2012. pp. 2780–3. [DOI] [PubMed] [Google Scholar]
- Mccarthy C, Feng D, Barnes N. Augmenting intensity to enhance scene structure in prosthetic vision. Electronic Proc. 2013 IEEE Int. Conf. on Multimedia and Expo Workshops, ICMEW 2013; 2013. [DOI] [Google Scholar]
- McCarthy C, Walker J G, Lieby P, Scott A, Barnes N. Mobility and low contrast trip hazard avoidance using augmented depth. J. Neural Eng. 2014;12:016003. doi: 10.1088/1741-2560/12/1/016003. [DOI] [PubMed] [Google Scholar]
- Meikle S J, Hagan M A, Price N S C, Wong Y T. Intracortical current steering shifts the location of evoked neural activity. J. Neural Eng. 2022;19:035003. doi: 10.1088/1741-2552/ac77bf. [DOI] [PubMed] [Google Scholar]
- Meikle S J, Wong Y T. Neurophysiological considerations for visual implants. Brain Struct. Funct. 2022;227:1523–43. doi: 10.1007/s00429-021-02417-2. [DOI] [PubMed] [Google Scholar]
- Mirochnik R M, Pezaris J S. Contemporary approaches to visual prostheses. Mil. Med. Res. 2019;6:19. doi: 10.1186/s40779-019-0206-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Moleirinho S, Whalen A J, Fried S I, Pezaris J S. The impact of synchronous versus asynchronous electrical stimulation in artificial vision. J. Neural Eng. 2021;18:051001. doi: 10.1088/1741-2552/abecf1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nakamara Y, Otsuka K. Motion discrimination as as test for visual motion perception. Nippon Ganka Gakkai Zasshi. 1995;99:1175–80. [PubMed] [Google Scholar]
- Naor O, Krupa S, Shoham S. Ultrasonic neuromodulation. J. Neural Eng. 2016;13:031003. doi: 10.1088/1741-2560/13/3/031003. [DOI] [PubMed] [Google Scholar]
- Nassi J J, Lyon D C, Callaway E M. The parvocellular LGN provides a robust disynaptic input to the visual motion area MT. Neuron. 2006;50:319–27. doi: 10.1016/j.neuron.2006.03.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nirenberg S, Pandarinath C. Retinal prosthetic strategy with the capacity to restore normal vision. Proc. Natl. Acad. Sci. U.S.A. 2012;109:15012–7. doi: 10.1073/pnas.1207035109. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Oswalt D, et al. Multi-electrode stimulation evokes consistent spatial patterns of phosphenes and improves phosphene mapping in blind subjects. Brain Stimul. 2021;14:1356–72. doi: 10.1016/j.brs.2021.08.024. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Palanker D, Le Mer Y, Mohand-Said S, Sahel J A. Simultaneous perception of prosthetic and natural vision in AMD patients. Nat. Commun. 2022;13:513. doi: 10.1038/s41467-022-28125-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Palanker D, Mer Y L, Mohand-Said S, Muqit M, Sahel J A. Photovoltaic restoration of central vision in atrophic age-related macular degeneration. Ophthalmology. 2020;127:1097–104. doi: 10.1016/j.ophtha.2020.02.024. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Paraskevoudi N, Pezaris J S. Eye movement compensation and spatial updating in visual prosthetics: mechanisms, limitations and future directions. Front. Syst. Neurosci. 2018;12:73. doi: 10.3389/fnsys.2018.00073. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Paraskevoudi N, Pezaris J S. Full gaze contingency provides better reading performance than head steering alone in a simulation of prosthetic vision. Sci. Rep. 2021;11:11121. doi: 10.1038/s41598-021-86996-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Parikh N, Itti L, Humayun M, Weiland J. Performance of visually guided tasks using simulated prosthetic vision and saliency-based cues. J. Neural Eng. 2013;10:026017. doi: 10.1088/1741-2560/10/2/026017. [DOI] [PubMed] [Google Scholar]
- Pavan A, Marotti R B, Mather G. Motion-form interactions beyond the motion integration level: evidence for interactions between orientation and optic flow signals. J. Vis. 2013;13:16. doi: 10.1167/13.6.16. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Perez-Yus A, Bermudez-Cameo J, Lopez-Nicolas G, Guerrero J. Depth and motion cues with phosphene patterns for prosthetic vision. 2017. [DOI]
- Perriello B. Retina Implant wins CE mark for next-gen blindness implant MassDevice. 2016. (available at: www.massdevice.com/retina-implant-wins-ce-mark-next-gen/)
- Petoe M A, et al. for the Bionics Institute and Centre for Eye Research Australia Retinal Prosthesis Consortium A second-generation (44-channel) suprachoroidal retinal prosthesis: interim clinical trial results. Transl. Vis. Sci. Technol. 2021;10:12. doi: 10.1167/tvst.10.10.12. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pezaris J S, Eskandar E N. Getting signals into the brain: visual prosthetics through thalamic microstimulation. Neurosurg. Focus. 2009;27:E6. doi: 10.3171/2009.4.FOCUS0986. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pezaris J S, Reid R C. Demonstration of artificial visual percepts generated through thalamic microstimulation. Proc. Natl Acad. Sci. USA. 2007;104:7670–5. doi: 10.1073/pnas.0608563104. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Picaud S, Sahel J-A. Retinal prostheses: clinical results and future challenges. C.R. Biol. 2014;337:214–22. doi: 10.1016/j.crvi.2014.01.001. [DOI] [PubMed] [Google Scholar]
- Ramachandran V S, Anstis S M. The perception of apparent motion. Sci. Am. 1986;254:102–9. doi: 10.1038/scientificamerican0686-102. [DOI] [PubMed] [Google Scholar]
- Ramirez K A, Drew-Bear L E, Vega-Garces M, Betancourt-Belandria H, Arevalo J F. An update on visual prosthesis. Int. J. Retina Vitr. 2023;9:73. doi: 10.1186/s40942-023-00498-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rassia K Eleonora, Moutoussis K, Pezaris J S. Reading text works better than watching videos to improve acuity in a simulation of artificial vision. Sci. Rep. 2022;12:12953. doi: 10.1038/s41598-022-10719-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rein D B, Wittenborn J S, Zhang P, Sublett F, Lamuda P A, Lundeen E A, Saaddine J. The economic burden of vision loss and blindness in the United States. Ophthalmology. 2022;129:369–78. doi: 10.1016/j.ophtha.2021.09.010. [DOI] [PubMed] [Google Scholar]
- Renshaw K T, Pezaris J S. Preprints.Org; 2025. Diversity of LGN-projecting primate retinal ganglion cells and their contributions to visual perception. [Google Scholar]
- Rizzo S, Belting C, Cinelli L, Allegrini L, Genovesi-Ebert F, Barca F, Di Bartolo E. The argus II Retinal prosthesis: 12-month outcomes from a single-study center. Am. J. Ophthalmol. 2014;157:1282–90. doi: 10.1016/j.ajo.2014.02.039. [DOI] [PubMed] [Google Scholar]
- Sadeghi R, Kartha A, Barry M P, Bradley C, Gibson P, Caspi A, Roy A, Dagnelie G. Glow in the dark: using a heat-sensitive camera for blind individuals with prosthetic vision. Vis. Res. 2021;184:23–29. doi: 10.1016/j.visres.2021.02.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sadeghi R, Kartha A, Barry M P, Gibson P, Caspi A, Roy A, Geruschat D R, Dagnelie G. Benefits of thermal and distance-filtered imaging for wayfinding with prosthetic vision. Sci. Rep. 2024;14:1313. doi: 10.1038/s41598-024-51798-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Salminen A-L, Karhula M E. Young persons with visual impairment: challenges of participation. Scand. J. Occup. Ther. 2014;21:267–76. doi: 10.3109/11038128.2014.899622. [DOI] [PubMed] [Google Scholar]
- Schaffrath K, et al. One-year safety and performance assessment of the Argus II retinal prosthesis: a postapproval study. JAMA Ophthalmol. 2019;137:896–902. doi: 10.1001/jamaophthalmol.2019.1476. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schiller P H. The effects of V4 and middle temporal (MT) area lesions on visual performance in the rhesus monkey. Vis. Neurosci. 1993;10:717–46. doi: 10.1017/S0952523800005423. [DOI] [PubMed] [Google Scholar]
- Schiller P H, Tehovnik E J. Visual prosthesis. Perception. 2008;37:1529–59. doi: 10.1068/p6100. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schmidt E M, Bak M J, Hambrecht F T, Kufta C V, O’Rourke D K, Vallabhanath P. Feasibility of a visual prosthesis for the blind based on intracortical micro stimulation of the visual cortex. Brain. 1996;119:507–22. doi: 10.1093/brain/119.2.507. [DOI] [PubMed] [Google Scholar]
- Seiffert A E, Somers D C, Dale A M, Tootell R B H. Functional MRI studies of human visual motion perception: texture, luminance, attention and after-effects. Cereb. Cortex. 2003;13:340–9. doi: 10.1093/cercor/13.4.340. [DOI] [PubMed] [Google Scholar]
- Shivdasani M N, Sinclair N C, Gillespie L N, Petoe M A, Titchener S A, Fallon J B, Perera T, Pardinas-Diaz D, Barnes N M, Blamey P J, for the Bionic Vision Australia Consortium Identification of characters and localization of images using direct multiple-electrode stimulation with a suprachoroidal retinal prosthesis. Invest. Ophthalmol. Vis. Sci. 2017;58:3962–74. doi: 10.1167/iovs.16-21311. [DOI] [PubMed] [Google Scholar]
- Sinclair N C, Shivdasani M N, Perera T, Gillespie L N, McDermott H J, Ayton L N, Blamey P J. for the Bionic Vision Australia Consortium The appearance of phosphenes elicited using a suprachoroidal retinal prosthesis. Invest. Ophthalmol. Vis. Sci. 2016;57:4948–61. doi: 10.1167/iovs.15-18991. [DOI] [PubMed] [Google Scholar]
- Solomon S G. Retinal ganglion cells and the magnocellular, parvocellular, and koniocellular subcortical visual pathways from the eye to the brain. In: Barton J J S, Leff A, editors. Handbook of Clinical Neurology. vol 178. Elsevier; 2021. pp. 31–50. ch 3. [DOI] [PubMed] [Google Scholar]
- Spencer T C, Fallon J B, Shivdasani M N. Creating virtual electrodes with 2D current steering. J. Neural Eng. 2018;15:035002. doi: 10.1088/1741-2552/aab1b8. [DOI] [PubMed] [Google Scholar]
- Stingl K, et al. Safety and efficacy of subretinal visual implants in humans: methodological aspects. Clin. Exp. Optom. 2013a;96:4–13. doi: 10.1111/j.1444-0938.2012.00816.x. [DOI] [PubMed] [Google Scholar]
- Stingl K, et al. Artificial vision with wirelessly powered subretinal electronic implant alpha-IMS. Proc. R. Soc. B. 2013b;280:20130077. doi: 10.1098/rspb.2013.0077. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stingl K, et al. Subretinal visual implant Alpha IMS—clinical trial interim report. Vis. Res. 2015;111:149–60. doi: 10.1016/j.visres.2015.03.001. [DOI] [PubMed] [Google Scholar]
- Stingl K, et al. Interim results of a multicenter trial with the new electronic subretinal implant Alpha AMS in 15 patients blind from inherited retinal degenerations. Front. Neurosci. 2017;11:445. doi: 10.3389/fnins.2017.00445. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stingl K, Bartz-Schmidt K-U, Gekeler F, Kusnyerik A, Sachs H, Zrenner E. Functional outcome in subretinal electronic implants depends on foveal eccentricity. Invest. Ophthalmol. Vis. Sci. 2013c;54:7658–65. doi: 10.1167/iovs.13-12835. [DOI] [PubMed] [Google Scholar]
- Strickland E, Harris M. Their bionic eyes are now obsolete and unsupported IEEE Spectrum. 2022. (available at: https://spectrum.ieee.org/bionic-eye-obsolete)
- Swenor B K, Lee M J, Varadaraj V, Whitson H E, Ramulu P Y. Aging with vision loss: a framework for assessing the impact of visual impairment on older adults. Gerontologist. 2020;60:989–95. doi: 10.1093/geront/gnz117. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Thomas L E, Lleras A. Moving eyes and moving thought: on the spatial compatibility between eye movements and cognition. Psychon. Bull. Rev. 2007;14:663–8. doi: 10.3758/BF03196818. [DOI] [PubMed] [Google Scholar]
- Titchener S A, et al. Oculomotor responses to dynamic stimuli in a 44-channel suprachoroidal retinal prosthesis. Transl. Vis. Sci. Technol. 2020;9:31. doi: 10.1167/tvst.9.13.31. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tootell R B H, Reppas J B, Dale A M, Look R B, Sereno M I, Malach R, Brady T J, Rosen B R. Visual motion aftereffect in human cortical area MT revealed by functional magnetic resonance imaging. Nature. 1995;375:139–41. doi: 10.1038/375139a0. [DOI] [PubMed] [Google Scholar]
- Turano K, Wang X. Motion thresholds in retinitis pigmentosa. Invest. Ophthalmol. Vis. Sci. 1992;33:2411–22. [PubMed] [Google Scholar]
- Van Essen D C, Maunsell J H R, Bixby J L. The middle temporal visual area in the macaque: myeloarchitecture, connections, functional properties and topographic organization. J. Comp. Neurol. 1981;199:293–326. doi: 10.1002/cne.901990302. [DOI] [PubMed] [Google Scholar]
- van Steveninck J R, Gestel T, Koenders P, Ham G, Vereecken F, Güçlü U, van Gerven M, Güçlütürk Y, Wezel R. Real-world indoor mobility with simulated prosthetic vision: the benefits and feasibility of contour-based scene simplification at different phosphene resolutions. J. Vis. 2022;22:1. doi: 10.1167/jov.22.2.1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- van Steveninck J R, Nipshagen M, van Gerven M, Güçlü U, Güçlüturk Y, van Wezel R. Gaze-contingent processing improves mobility, scene recognition and visual search in simulated head-steered prosthetic vision. J. Neural Eng. 2024;21:026037. doi: 10.1088/1741-2552/ad357d. [DOI] [PubMed] [Google Scholar]
- Veraart C, Wanet-Defalque M-C, Gérard B, Vanlierde A, Delbeke J. Pattern recognition with the optic nerve visual prosthesis. Artif. Organs. 2003;27:996–1004. doi: 10.1046/j.1525-1594.2003.07305.x. [DOI] [PubMed] [Google Scholar]
- Villalobos J, et al. A wide-field suprachoroidal retinal prosthesis is stable and well tolerated following chronic implantation. Invest. Ophthalmol. Vis. Sci. 2013;54:3751–62. doi: 10.1167/iovs.12-10843. [DOI] [PubMed] [Google Scholar]
- Volta A. XVII. On the electricity excited by the mere contact of conducting substances of different kinds. In a letter from Mr. Alexander Volta, F. R. S. Professor of Natural Philosophy in the University of Pavia, to the Rt. Hon. Sir Joseph Banks, Bart. K.B. P. R. S. Phil. Trans. R. Soc. 1997;90:403–31. doi: 10.1098/rstl.1800.0018. [DOI] [Google Scholar]
- Wallach H. Eye Movement and Motion Perception. In: Wertheim A H, Wagenaar W A, Leibowitz H W, editors. Tutorials on Motion Perception. Springer; 1982. pp. 1–18. [DOI] [Google Scholar]
- Wang J, Lu Y, Gu L, Zhou C, Chai X. Moving object recognition under simulated prosthetic vision using background-subtraction-based image processing strategies. Inf. Sci. 2014;277:512–24. doi: 10.1016/j.ins.2014.02.136. [DOI] [Google Scholar]
- Wang J, Zhu H, Liu J, Li H, Han Y, Zhou R, Zhang Y. The application of computer vision to visual prosthesis. Artif. Organs. 2021;45:1141–54. doi: 10.1111/aor.14022. [DOI] [PubMed] [Google Scholar]
- Wertheimer M. Experimentelle Studien über das Sehen von Bewegung. Z. Psychol. Physiol. Sinnesorg. 1912;61:161–5. [Google Scholar]
- Wilke R, et al. Spatial resolution and perception of patterns mediated by a subretinal 16-electrode array in patients blinded by hereditary retinal dystrophies. Invest. Ophthalmol. Vis. Sci. 2011;52:5995–6003. doi: 10.1167/iovs.10-6946. [DOI] [PubMed] [Google Scholar]
- Yamasaki D S, Wurtz R H. Recovery of function after lesions in the superior temporal sulcus in the monkey. J. Neurophysiol. 1991;66:651–73. doi: 10.1152/jn.1991.66.3.651. [DOI] [PubMed] [Google Scholar]
- Yan Y, Lu Y, Li M, Ma Z, Cao P, Chen Y, Sun X, Chai X, Ren Q, Li L. Electrically evoked responses in the rabbit cortex induced by current steering with penetrating optic nerve electrodes. Invest. Ophthalmol. Vis. Sci. 2016;57:6327–38. doi: 10.1167/iovs.15-17543. [DOI] [PubMed] [Google Scholar]
- Yanai D, Weiland J D, Mahadevappa M, Greenberg R J, Fine I, Humayun M S. Visual performance using a retinal prosthesis in three subjects with retinitis pigmentosa. Am. J. Ophthalmol. 2007;143:820–7.e2. doi: 10.1016/j.ajo.2007.01.027. [DOI] [PubMed] [Google Scholar]
- Yoon Y H, Humayun M S, Kim Y J. One-year anatomical and functional outcomes of the argus ii implantation in Korean patients with late-stage retinitis pigmentosa: a prospective case series study. Ophthalmologica. 2021;244:291–300. doi: 10.1159/000513585. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yue L, Falabella P, Christopher P, Wuyyuru V, Dorn J, Schor P, Greenberg R J, Weiland J D, Humayun M S. Ten-year follow-up of a blind patient chronically implanted with epiretinal prosthesis Argus I. Ophthalmology. 2015;122:2545–52.e1. doi: 10.1016/j.ophtha.2015.08.008. [DOI] [PubMed] [Google Scholar]
- Zavala A. Development of the forced-choice rating scale technique. Psychol. Bull. 1965;63:117–24. doi: 10.1037/h0021567. [DOI] [PubMed] [Google Scholar]
- Zheng D D, Swenor B K, Christ S L, West S K, Lam B L, Lee D J. Longitudinal associations between visual impairment and cognitive functioning: the salisbury eye evaluation study. JAMA Ophthalmol. 2018;136:989–95. doi: 10.1001/jamaophthalmol.2018.2493. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
No new data were created or analysed in this study.





