Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2011 Feb 1.
Published in final edited form as: Phys Med Rehabil Clin N Am. 2010 Feb;21(1):15–32. doi: 10.1016/j.pmr.2009.07.006

RESEARCH IN COMPUTER ACCESS ASSESSMENT AND INTERVENTION

Richard Simpson 1,2,, Heidi Horstmann Koester 3, Edmund LoPresti 3,4
PMCID: PMC2788505  NIHMSID: NIHMS136407  PMID: 19951775

1 IMPORTANCE OF COMPUTER ACCESS TECHNOLOGY

Computer Access Technology (CAT) allows people who have trouble using a standard computer keyboard, mouse or monitor to access the computer. CAT includes relatively inexpensive devices like trackballs and small-footprint keyboards as well as sophisticated technologies like automatic speech recognition, eyegaze tracking, and brain-computer interfaces. CAT services are provided by a range of rehabilitation professionals, including rehabilitation engineers, occupational therapists, speech-language pathologists, special educators and vocational rehabilitation counselors.

CAT is critical for enhancing the educational and vocational opportunities of people with disabilities [1, 2]. In addition, CAT has been shown to contribute to improved health status by providing access to health information and interaction with clinicians and peers [3]. CAT can reduce social isolation by eliminating physical barriers, facilitating communication, and providing a forum for the exchange of information [4]. Individuals with disabilities often appreciate the anonymity of the Internet, where they can be evaluated for the strength of their contributions rather than their physical appearance or disability [5, 6]. The Internet also provides protection against self-consciousness and social anxiety, and active participation can lead to greater levels of self-acceptance and decreased feelings of isolation [7, 8].

57% (74.2 million) of working-age (between 18 and 64 years of age) computer users are likely or very likely to benefit from the use of CAT [9]:

  • 17% (21.9 million) of working-age computer users have a mild visual difficulty or impairment, and 9% (11.1 million) have a severe visual difficulty or impairment [9].

  • 19% (24.4 million) of working-age computer users have a mild dexterity difficulty or impairment, and 5% (6.8 million) have a severe dexterity difficulty or impairment [9].

  • 18% (24.0 million) of working-age computer users have a mild hearing difficulty or impairment, and 2% (2.5 million) have a severe hearing difficulty or impairment [9].

Choosing the most appropriate CAT is a collaborative decision-making process involving the consumer, clinician(s), and third-party payers. The challenges involved in a successful computer access intervention include:

  1. Evaluating and documenting client abilities and specific difficulties with the standard computer interface [10];

  2. Choosing the most appropriate assistive technology to address these difficulties;

  3. Configuring the technology to the user’s needs;

  4. Training the user in appropriate use of their system [1113]; and

  5. Providing continuous follow-up to ensure that the interface remains well-suited to the user [14, 15, 13, 16, 17].

The consequences of failing to successfully meet any one of these challenges include wasted human and material resources spent in the intervention process, unnecessary obstacles placed in the individual’s path toward their goals, technology abandonment [18, 19] and a lack of computer use amongst individuals with disabilities [3, 20]. By all reported measures, we are largely failing to meet these challenges.

People with disabilities do not use computers at the same rate as people without disabilities. While 85% of working-age adults without disabilities use computers, computer usage is 80% among those who have mild impairments and is 63% among individuals who have severe disabilities and are very likely to benefit from the use of CAT [9].

Many people with disabilities who do own computers do not take advantage of CAT. One study found that only 24% of working-age computer users with severe disabilities use CAT [21]. A significant barrier to obtaining CAT is cost. 28% of working-age computer users with mild difficulties/impairments and 32% with severe disabilities report there is a CAT that they do not own but would purchase if it became more affordable [21].

A second obstacle is awareness. Of computer users with mild or severe disabilities, 20% were not aware that CAT existed which was appropriate for them [21]. Awareness of the free, built-in accessibility settings offered by the Windows and Mac operating systems is even lower. Of computer users with mild or severe disabilities, 36% were unaware of the available adjustments for the mouse and 35% were unaware of the available keyboard adjustment options [21].

Amongst computer users who do own and use CAT, there is ample evidence that it is not configured correctly. For example, Trewin and Pain reported target acquisition error rates of greater than 10% for 14 of 20 users with physical disabilities, and observed that 55% of the dragging tasks made by these users were unsuccessful [22]. An average of 28% of clicks in this study included a mouse movement, which is a potential source of error, and 40% of multiple click attempts were unsuccessful [22].

Worst of all, approximately a third of computer users who do receive CAT wind up abandoning it. A study of 115 individuals with disabilities who received 136 assistive technology devices over five years, including computers, communication devices and adapted software, reported a total abandonment rate of 32.4%. The abandonment rate within the study for computer access and communication devices was 30.8% [23].

2 CHALLENGES TO COMPUTER ACCESS SERVICE DELIVERY

2.1 THE COST OF CAT AND CLINICAL SERVICES

A significant obstacle to obtaining the most appropriate CAT is cost. CAT (for working-age adults) is often funded through vocational rehabilitation agencies which have small budgets relative to the demand for their services. Funds spent on clinician time and CAT for one client are unavailable to other clients, so maximizing each dollar spent is critical. Consumers who are retired, or not seeking employment, are equally cost sensitive as they may have to purchase their own equipment. As the population ages, an increasing number of older computer users will potentially benefit from CAT [21] but are unlikely to be candidates for vocational rehabilitation and are thus likely to be using their own funds.

2.2 THE VARIETY OF DEVICES AVAILABLE

Another challenge for computer access is the simple fact that there are too many options for any consumer to reasonably compare. There is a nearly limitless variety of alternatives to the traditional computer keyboard and mouse. Within each of these categories of devices there are multiple products, and each product has its own unique set of configuration options.

Given this variety of options, a consumer could spend weeks or months comparing devices, but a decision is often needed after one or two clinical sessions. In particular, consumers who need to obtain computer access technology in order to return to work or school often need to make decisions quickly to meet other deadlines. Beyond that, most consumers lack the patience or funding for multiple clinical visits. It is therefore critical to make optimal use of the time that client and clinician spend together in the clinic.

2.3 CONSUMER NEEDS CHANGE OVER TIME

Even if a consumer who needs CAT receives a successful technology intervention, it is unlikely that the solution will remain effective indefinitely. Indeed, a consumer’s needs can change long before he or she has the opportunity to be re-evaluated by a clinician. Regular follow-up evaluations are critically important [1113], but often do not occur due to a variety of barriers, such as the time and travel required and a lack of funding.

In addition, a single configuration may not be appropriate for a user at all times. A user’s needs may change due to changes in his or her abilities which may happen over the course of a day (e.g., fatigue) or longer (e.g., due to progression of the disability, recovery of function, or other factors). The user’s needs may also change based on the user’s desired tasks (e.g., some computer activities may require greater precision than others). Even if a clinician is available to recommend an initial configuration, he or she may not be available every time adjustments to the configuration are desirable. If a user is responsible for his or her own adjustments, he or she may not notice, or know how to respond to, a gradual decline in performance.

3 POTENTIAL TECHNOLOGY SOLUTIONS

The problems with the computer access assessment process will not be solved through technology alone. The assessment process should be based on each client’s individual priorities, taking into account preferences, physical and functional needs, social environments, and related issues [24], which makes the involvement of a trained clinician crucial. A clinician will always be needed to work with the consumer and other stakeholders to choose the most appropriate technology. Similarly, a clinician or other caregiver is irreplaceable for training in many circumstances. Likewise, funding and reimbursement issues are public policy problems rather than technology problems.

In other areas, however, technology can play a prominent supporting (or even leading) role. These include computer access assessment tools that clinicians can use to evaluate and document a client’s abilities, utilities to automatically configure technology for each client’s individual needs, and technology to provide follow-up services through remote client-clinician interaction. Research in each of these areas is discussed below.

3.1 ASSESSMENT TOOLS

Many different methods are currently employed in assessments for computer access. Informal clinical observation is the oldest method and is characterized by brief trials with candidate systems in a clinic setting and qualitative judgments of client performance during these trials. The lack of an explicit, objective framework in this approach makes it likely that appropriate candidate systems may not be considered, that criteria for selecting among candidate systems may not be clearly defined, and that objective data to guide the ultimate selection may not be available, particularly when evaluations are carried out by relatively inexperienced clinicians.

The need for a systematic approach to delivering assistive technology services has long been recognized (e.g., [25, 26]). The systems and approaches developed to date are diverse and can be categorized on the following dimensions:

  1. types of information gathered (qualitative vs. quantitative);

  2. method of data collection and management (manual vs. computerized);

  3. focus of approach (assessment of ability vs. prescription of “best” device).

Several conceptual models for the “ideal” assistive technology assessment process have been developed. These conceptual models are designed to structure the evaluation to ensure that important considerations are not overlooked. Forms and worksheets are often provided to help guide the evaluator’s activities and lead to a recommendation of the most appropriate device. A weakness in this approach is that little or no support is provided for measuring specific aspects of client performance, such as speed and accuracy. The procurement and management of objective quantitative data are left to the evaluator, who may or may not perform this task in a consistent or valid way.

Generalized qualitative assessment models (applying to all aspects of assistive technology) include:

  • The Student, Environment, Tasks, Tools (SETT) model [27]. A general framework (rather than an actual protocol) for evaluating a person, environment and goals to identify the most appropriate assistive technology.

  • Matching Persons with Technology (MPT) [28, 13, 29, 30]. A validated measure that evaluates how individuals judge their own functional and health status. The MPT measures quality of life and predisposition towards AT use, and has been shown to successfully predict satisfaction with AT one month after discharge [29].

  • Considering Assistive Technology [31]. A flowchart to guide professionals through consideration of assistive technology by asking a series of questions.

  • Assessing Students’ Needs for Assistive Technology [32]. A protocol for evaluating a child’s assistive technology needs in the context of an Individual Education Plan (IEP).

  • Education Tech Points [33]. Assessment forms and a manual documenting the components of effective AT service delivery.

As shown in

Tool Recommends
Assessment
Protocol
Computerized
Skill Evaluation
Automatic
Report
Generation
Recommends
Devices
Skills Evaluated
Compass N Y Y N Use of text entry pointing devices; sensory abilities; cognitive abilities
REACH Interface Author N Y Y N Use of text entry and pointing devices; switch skills
Alternative Computer Access: A Guide to Selection Y N N N Use of text entry and pointing devices; sensory abilities; cognitive abilities
Control of Computer-Based Technology for People with Disabilities Y N N N Background; environment; text entry, pointing and switch skills
Lifespace Y N Y N Physical, cognitive, emotional, support resources and environmental characteristics
Assessment of Computer Task Performance Y N N N Text entry and pointing skills
EvaluWare N Y Y N Looking, listening, pointing, switch, some text entry skills
VOCAselect Y N Y Y Criteria for an appropriate AAC device
Computer Access Selector Y N Y Y Criteria for an appropriate computer access device

Table 1, several assessment tools have been developed specifically for computer access. Qualitative assessment methods designed specifically for computer access include:

  • Alternative Computer Access: A Guide to Selection [34]. A decision tree which guides clinicians through the assessment process. Decision points prompt the clinician to evaluate the client’s motor, sensory and cognitive skills. The “leaves” of the tree are suggestions of appropriate types of computer access devices.

  • Control of Computer-Based Technology for People with Physical Disabilities: An Assessment Manual [35]. A manual with data collection instruments, procedures for client observations, testing procedures for various computer access methods, and guidelines for matching client needs and device characteristics.

  • Assessment of Computer Task Performance [36]. A series of tests and measurement criteria for assessing computer skills. No software is provided to support testing. Instead, the tests are administered using icons present on the computer desktop and standard word processing software. A series of transparencies are provided which can be placed on the computer monitor to support the tests (e.g., indicating a path which the client should follow with the cursor).

  • Lifespace Access Profile [37]. A team-based assessment tool that uses worksheets to measure physical, perceptual, cognitive and emotional skills, support resources, and environmental considerations. The worksheets are computerized, but data consists of subjective ratings from each team member.

  • EvaluWare. A software package that presents evaluation exercises for a range of computer access skills. Despite the use of computerized tests, EvaluWare does not automatically record performance data.

Table 1.

Computer access assessment models and products

Tool Recommends Assessment Protocol Computerized Skill Evaluation Automatic Report Generation Recommends Devices Skills Evaluated
Compass N Y Y N Use of text entry pointing devices; sensory abilities; cognitive abilities
REACH Interface Author N Y Y N Use of text entry and pointing devices; switch skills
Alternative Computer Access: A Guide to Selection Y N N N Use of text entry and pointing devices; sensory abilities; cognitive abilities
Control of Computer-Based Technology for People with Disabilities Y N N N Background; environment; text entry, pointing and switch skills
Lifespace Y N Y N Physical, cognitive, emotional, support resources and environmental characteristics
Assessment of Computer Task Performance Y N N N Text entry and pointing skills
EvaluWare N Y Y N Looking, listening, pointing, switch, some text entry skills
VOCAselect Y N Y Y Criteria for an appropriate AAC device
Computer Access Selector Y N Y Y Criteria for an appropriate computer access device

Several computerized systems attempt to support the evaluator through all stages of the assessment process, producing a recommendation for the most appropriate assistive device. Examples of these systems include:

  • Computer Access Selector [38]. Uses device criteria chosen from a series of prompts by a clinician to identify the most appropriate device from a list of known devices.

  • Assistive Technology Expert System [39, 40]. Presents a series of questions to the clinician based on rules in an expert system. Answers to each question determine subsequent questions, until a device is identified.

Systems with a primary focus on device prescription generally use assessment data only as a means to that end, which tends to limit the main use for these systems to “one-time” major evaluations lasting several hours or more. They are not easily applicable to assessments that take place in a single therapy session and are not designed to track performance across multiple assessments. Furthermore, the systems developed to date are limited to recommending a single device at a time, and are unable to coordinate a multi-faceted approach to computer access.

What most models, protocols, guidelines and tools uniformly lack is a means of collecting detailed quantitative performance data for use in decision making. As a result, computer access assessments primarily consist of brief trials with candidate systems resulting in qualitative judgments of client performance. Some clinicians collect quantitative data with a stopwatch, typing test software, or video games, but these approaches do not necessarily produce valid comparisons between devices.

The Compass software system [4144], which measures users’ skills in various kinds of computer interaction, is designed to help clinical and educational professionals perform computer access evaluations with their clients by providing them with a clear picture of a client’s strengths and limitations. Compass tests the skill families of text entry, pointing and switch use through a hierarchy of tests which tap into successively more complex aspects of each skill. A hierarchy of complexity helps accommodate differing client abilities. For example, matching single letters of the alphabet may be a more appropriate assessment of keyboarding skill in a young student than transcription of full sentences. The hierarchy also provides a way to isolate the physical component of the test from its perceptual and cognitive aspects. As one moves up the hierarchy, tests incorporate more perceptual and cognitive skills. Performance on higher level tests may be compared to performance at the lower level of the hierarchy to reveal how perceptual and cognitive issues affect keyboarding and pointing for a particular client.

3.2 AUTOMATICALLY ADAPTING DEVICE CONFIGURATION

The behavior of most computer input devices, such as keyboards and mice, is adjustable. Because each person’s disability is unique, tuning these devices to a user’s strengths and limitations is critical for success in many cases. Ideally, configuration is performed in consultation with a clinician who has expertise in computer access for people with disabilities. However, a trained clinician may not be available, and even when one is, proper tuning of a device to the needs of a particular user can be a difficult and time- consuming task. The challenge is magnified by the fact that user needs and abilities may change over time, whether in the short term due to factors such as fatigue or in the long term due to factors such as changes in the individual’s underlying impairment. For these reasons, input devices are often not appropriately configured to meet users' needs, with consequent negative effects on user productivity and comfort.

A user’s system is typically configured in one or more of three ways. The first, and perhaps most common, is to use the default values for the device. Moderately inappropriate values may result in multiple keyboarding errors and/or difficulty selecting targets with the mouse, decreasing user performance and satisfaction. In a more extreme case, the system may be virtually unusable under the default values.

A second scenario is when the user makes his or her own adjustments. This requires that the user knows what parameters are available and how to adjust them. This is a complex task. Performing all possible adjustments for keyboard and mouse within Windows XP requires accessing three separate Control Panel applications with 12 tabbed panels, while making objects larger for easier selection would require accessing a number of additional Control Panel applications. Terminology can be ambiguous; for example, to invoke BounceKeys, the user chooses to “ignore repeated keystrokes,” while to adjust the repeat settings, the user must select “ignore quick keystrokes.” Another potential source of confusion is that the repeat settings can be adjusted in two different control panels, with the accessibility settings overriding the keyboard control panel settings. Even if the user can successfully navigate the options, knowing the most appropriate values for all applicable settings may be even more difficult. Users may not understand how the parameter settings relate to the interface problems they are having, or if they do, the best choice of specific values may be unclear. Recent versions of the Windows operating system include an accessory program called the Accessibility Wizard, which does provide some help in reducing the complexity of configuration for keyboards, pointers, and the visual display. However, it does not include all available settings (e.g., the key repeat settings are not available), nor does it give specific suggestions about how to appropriately set parameter values based on user performance.

A third scenario occurs when a clinician or teacher is available to assist with the configuration process, using clinical observations and knowledge of the possible accommodations as a guide. However, most users with physical disabilities do not have a qualified clinician available to them. Trewin and Pain [45] found that only 35% of 30 computer users with physical disabilities had a “computer teacher.” Further, not all clinicians have the skills to effectively assist. Even when a clinician or other advisor with relevant expertise is available, input device configuration often requires considerable trial and error.

Under each of these three approaches, it may be difficult to define appropriate settings for a user’s initial configuration. It is equally difficult, if not more so, to address changes in the user’s abilities over time, which may happen over the course of a day, a month, or a year, depending on the nature of the user’s disability. Current methods may lead to appropriate input device configurations in some cases, but it does take special knowledge, additional time, and continued maintenance to do it right [34]. As a result, input devices are often not appropriately configured to meet users’ needs, with consequent negative effects on user productivity and comfort.

An automated agent on the user’s computer could help ensure that input devices are properly configured for the individual, and reconfigured as the user’s needs change. Such an agent would need a means to observe the user’s performance and predict appropriate input device configuration settings based on that performance. Several groups have been working toward configuration agents that would support this process [4651]. A configuration agent models a user’s strengths and limitations, and based on the model, helps configure the user’s input devices appropriately.

3.2.1 Automatically Configuring Keyboards

Tuning a keyboard to a user’s strengths and limitations may yield significant performance and comfort benefits. Conversely, the potential consequences of inappropriate settings are many [1]. For example, for someone who types with a mouthstick, not having StickyKeys active makes it cumbersome to type capital letters and impossible to use other key combinations such as Ctrl-C.

Trewin and colleagues have been developing a configuration agent for keyboard settings [5257, 22]. The agent creates a user model based on free typing and determines settings for a range of parameters such as StickyKeys, Repeat Delay, and BounceKeys (see

Tool Recommends
Assessment
Protocol
Computerized
Skill Evaluation
Automatic
Report
Generation
Recommends
Devices
Skills Evaluated
Compass N Y Y N Use of text entry pointing devices; sensory abilities; cognitive abilities
REACH Interface Author N Y Y N Use of text entry and pointing devices; switch skills
Alternative Computer Access: A Guide to Selection Y N N N Use of text entry and pointing devices; sensory abilities; cognitive abilities
Control of Computer-Based Technology for People with Disabilities Y N N N Background; environment; text entry, pointing and switch skills
Lifespace Y N Y N Physical, cognitive, emotional, support resources and environmental characteristics
Assessment of Computer Task Performance Y N N N Text entry and pointing skills
EvaluWare N Y Y N Looking, listening, pointing, switch, some text entry skills
VOCAselect Y N Y Y Criteria for an appropriate AAC device
Computer Access Selector Y N Y Y Criteria for an appropriate computer access device

Table 1). The agent’s recommendations were evaluated with 20 keyboard users who have physical disabilities. For StickyKeys, the agent’s recommendation correlated significantly with users’ opinions on how useful StickyKeys would be for them. However, the discrimination of the agent was imperfect, as 9 users felt that StickyKeys was useful for them, even though the agent did not recommend it for them. For repeat delay, use of the agent-recommended settings significantly reduced key repeat errors (from 2610 to 151 errors) [58]. The agent accurately recommended use of BounceKeys for 5 of 7 subjects who made bounce errors. Effects on productivity measures, such as typing speed, were not measured.

One of the challenges in Trewin’s approach is that it makes inferences based on unconstrained typing tasks. The difficulty of this approach is shown in the fact that the agent accurately detected only 55% of inadvertent keypress errors [45] The use of unconstrained typing tasks allows for continuous monitoring, which is less obtrusive to the user, but may compromise the success of the agent’s suggestions.

Koester, LoPresti and Simpson are developing a software system called IDA (Input Device Agent), whose goal is to optimally configure input devices for people with physical impairments [59]. In a study of twelve typists with physical impairments, IDA recommended three keyboard parameters in response to measurements of typing performance: repeat rate, repeat delay, and use of StickyKeys. For two participants with significant problems with inadvertent key repeats, use of the IDA-recommended repeat settings reduced the number of repeated characters by 96% and significantly improved text entry rate and typing accuracy. IDA recommended StickyKeys for six participants, which eliminated their modifier-related errors and significantly improved their typing speed. IDA did not recommend StickyKeys for the six participants who demonstrated no need for it.

3.2.2 Automatically Configuring Pointing Devices

Pointing devices include the standard mouse as well as trackballs, laptop trackpads, head-controlled pointers, and many other devices. Typical parameters for pointing device configuration include those shown in

Parameter Description
Button-handedness Controls the functions assigned to the left and right mouse buttons.
Click method Whether the user performs a single or double click to select icons.
Double-click speed Controls the allowable time between two clicks in a double-click.
Pointer speed (Gain) How quickly the cursor moves across the screen in response to mouse movements.
Enhanced pointer precision The Enhance Pointer Precision (EPP) setting enables a complex algorithm controlling the velocity and acceleration of the mouse cursor.
Snap-to-default If this option is active, when a dialog box appears on the screen, the cursor will immediately move to the default button (e.g., “OK”).
Object Size It is possible to change the size (in pixels) of icons, menu bars, and other objects in the user interface. Increasing the size of these objects may make them easier to select, at the cost of reduced space on the screen. There may be many separate settings which could be adjusted to make different classes of objects larger.

Table 3. Other settings may be available, depending on the specific device and device driver being used. For example, Logitech trackballs (Logitech, Fremont, CA) allow the user to program the buttons to perform different functions. The trackpads common on laptop computers include settings related to cursor speed and whether the user can click by pressing on the trackpad or whether he or she must use the buttons.

Table 3.

Typical configuration options for pointing devices.

Parameter Description
Button-handedness Controls the functions assigned to the left and right mouse buttons.
Click method Whether the user performs a single or double click to select icons.
Double-click speed Controls the allowable time between two clicks in a double-click.
Pointer speed (Gain) How quickly the cursor moves across the screen in response to mouse movements.
Enhanced pointer precision The Enhance Pointer Precision (EPP) setting enables a complex algorithm controlling the velocity and acceleration of the mouse cursor.
Snap-to-default If this option is active, when a dialog box appears on the screen, the cursor will immediately move to the default button (e.g., “OK”).
Object Size It is possible to change the size (in pixels) of icons, menu bars, and other objects in the user interface. Increasing the size of these objects may make them easier to select, at the cost of reduced space on the screen. There may be many separate settings which could be adjusted to make different classes of objects larger.

Proper adjustment of these settings can be critical to efficient use of the pointing device for people with disabilities [1]. For someone with impaired motor control, the default pointer speed on the pointing device may cause the cursor to move much too quickly, making it difficult or impossible to select small targets such as toolbar buttons. Other difficult tasks include dragging the pointer with the mouse button depressed and clicking and double-clicking the mouse button while keeping the pointer still.

If a relationship can be found between an individual’s movement patterns and his or her optimal configuration settings, a software agent could customize the settings in response to the user’s needs. This concept has been explored with a force-sensing joystick which adapted to hand tremor using measurements of the user’s tracking ability and tremor [48]. A preliminary study with three subjects who had Friedrich’s ataxia indicated that the adaptive joystick provided some improvement in performance for tracking tasks.

Tracey and Winters developed a system to configure mouse settings in the Windows operating system based on subject performance on computer tracking exercises as well as direct questions directed toward an assistive technology clinician who had observed the user [12]. The principle limitation of this system is its assumption of the presence of an assistive technology expert when, in some cases, users may need or want to configure their systems independently

LoPresti developed and evaluated a system that automatically adjusted the gain for users of head-controlled pointing devices [17]. For 16 subjects with physical disabilities, the system was able to select settings that were appropriate for most subjects and provided a modest but significant improvement in performance (p < 0.05). IDA recommends a setting for the computer’s control-display gain based on observations of a user’s performance in a target selection task [60]. In a study involving 12 participants who have motor impairments, the IDA-selected gain was not associated with significant improvements in selection time or error-free performance compared with the operating system’s default gain. However, two participants did have notable and consistent improvement in selection time and error-free performance using the IDA-selected gain [60].

3.2.3 Switches

Individuals who cannot use adapted keyboards or pointing devices may use switch-based input techniques. One example is single-switch scanning, in which the system presents choices sequentially to the user. A common implementation of single-switch scanning requires three switch hits to make one selection from a row-column matrix of letters, numbers, symbols, words, or phrases. The first switch hit initiates a scan through the rows of the matrix. Each row of the matrix is highlighted in turn until a second switch hit is made to select a row. Each column of the row is then highlighted in turn until the target is highlighted, when the third switch hit is made to select the target. Depending on the scanning system used, there may be three or more adjustable parameters (see

Parameter Description
Scan period The amount of time an item remains highlighted for the user to make a selection
Initial Scan Delay Additional delay applied to the first row or column
Column Scans Maximum number of times the columns within a row are scanned
Layout Arrangement of targets within the scanning matrix

Table 4), with the scan rate being most important. Switch-based parameters are incorporated into the switch system used and are not a feature of the operating system. The scan rate and other configuration parameters determine the minimum letter selection time which is possible for a user. If the scan rate in single-switch scanning is too fast, the user will make a lot of errors or may be unable to use the system. A scan rate which is too slow will unnecessarily slow down performance in a process that is already inherently very slow [1].

Table 4.

Typical configuration options for single-switch scanning.

Parameter Description
Scan period The amount of time an item remains highlighted for the user to make a selection
Initial Scan Delay Additional delay applied to the first row or column
Column Scans Maximum number of times the columns within a row are scanned
Layout Arrangement of targets within the scanning matrix

One-switch row-column scanning can be tiring to use and is generally a relatively slow method of communication. An able-bodied individual using an optimally-designed matrix of 26 letters and a space can produce between 6 and 8 words/minute using this method [61, 62]. Despite its limitations, however, row-column scanning fills an important niche within access techniques by providing an affordable and reliable option for many individuals with limited movement and limited vocal abilities. Hence, despite increasing interest in speech recognition, eye-tracking, and direct-brain interfaces for accessing assistive technology, there remain valid reasons for seeking to enhance performance using row-column scanning.

Three research groups have worked on methods of automatically adapting the scan period of a single-switch row-column scanning system. Cronk and Schubert [63] developed an expert system for the adaptation of scan period, but it was never integrated into any commercial systems. Lesher et al. [64, 65] developed a rule-based method of scan period adjustment based on user errors and the time required for the user to make a selection relative to the available time. Their primary goal was to provide a means of scan period adjustment for empirical studies comparing different scanning displays, and their system performed well enough to meet this goal with able-bodied subjects [65].

Simpson and Koester [66] developed and evaluated a single switch scanning system that used a Bayesian network to adjust the user’s scan period in real time. Two studies, involving a total of 16 subjects without disabilities, demonstrated that the system could make reasonable adaptation decisions, with no human intervention, for a system with a single scan delay. Subjects’ text entry performance and subjective opinion was no different with the automatic system as compared to a manual adjustment protocol, in which able-bodied subjects could change the scan period at will with a single keypress. A major limitation is that the work was not validated with users with disabilities.

IDA’s ability to make recommendations for scan period in a row-column scanning system was evaluate with two groups of individuals (8 people who were either able-bodied or had spinal cord injuries and 6 individuals with severe physical disability secondary to cerebral palsy) [67]. Participants’ speed, accuracy, and subjective ratings when using the IDA-recommended scan rate suggest that IDA can recommend an appropriate scan rate. Participants’ performance was at least as good for the IDA-selected scan period as for the self-selected scan period [67].

3.3 MONITORING AND TELEREHABILITATION

The Institute of Medicine defines telemedicine as “the use of electronic information and communications technologies to provide and support health care when distance separates the participants [68].” The Shepherd Center has further defined telerehabilitation as “the use of telecommunications technology to provide rehabilitation and long-term support to people with disabilities [69].” Telerehabilitation technology can either be used interactively or in a store-and-forward mode. Interactive telerehabilitation sessions involve both the client and clinician participating in the session at the same time, typically using audio- and video-conferencing technology. Store-and-forward telerehabilitation sessions, on the other hand, allow the clinician to collect data from the client without the need for simultaneous interaction.

Telerehabilitation technology can allow a clinician to collect performance data during a “loan period” to evaluate how well a potential computer access solution meets the consumer’s needs. Quite often, the sole measurable outcome of a loan period is the consumer’s subjective impression of the AT. While the consumer’s impression is important, there is typically no record of how often the device was used during the loan period, how well the consumer performed when using the device, or how the device was configured - all of which is important for making informed comparisons between multiple potential solutions. More importantly, self-reports can be unreliable. Studies [7072] have shown that “estimates of concordance of client self-reports and proxy-reports with performance-based measures range from a low of 63% to a high of 94% depending on the specific daily living task [73].”

Telerehabilitation can also be a cost-effective solution to providing follow-up to consumers after a device has been purchased. Receiving a device is no guarantee that the device will be installed and configured correctly, particularly if funding for an in-person visit from a clinician is not available. Telerehabilitation may reduce the cost of a clinician’s involvement in the installation and configuration process to the point where this service can be provided with each piece of equipment received. Follow-up assistance may take several forms, including:

  • interactive desktop sharing, video conferencing or text messaging to allow the clinician to guide (or control) the installation/configuration process;

  • interactive or store-and-forward use of skill tests, which allow the clinician to compare performance obtained in the clinic with results on the consumer’s home or work computer.

Several clinical teams have documented the use of telerehabilitation to enhance the provision of computer access and augmentative communication services, although these publications have thus far been limited to case studies with little or no statistical analysis. In the United States, the Shepherd Center of Georgia has perhaps the most experience using telerehabilitation technology for computer access. They have described one instance in which a videophone was used to evaluate how well a computer access system was configured for a client [69], and have also described using telerehabilitation technology for training a client to use speech-recognition software for computer access [74].

Another American researcher, Elliot Cole, has described a computer-based cognitive rehabilitation intervention that made use of telecommunications technology [75]. Cole developed computer software that assisted a client with communication and information storage and retrieval. The client was trained to use the software using videoconferencing technology and desktop sharing software. Perhaps more importantly, this is the only reported instance we could identify in which telerehabilitation was used as a means of conducting follow-up evaluations. Specifically, Cole modified the system based on data automatically collected by the system during regular use.

In Europe, the Oxford Aiding Communication in Education (ACE) Center has made extensive use of telerehabilitation technology to support computer access and augmentative communication interventions. The Telenet [76] and CATCHNET [16] projects evaluated the effectiveness of low-cost videoconferencing and online software sharing in providing computer access assessment, training and support. Both projects used an ISDN line to connect clinicians at the ACE Center with clinicians and clients in remote clinical sites, and reported general client and clinician satisfaction with the approach.

The Telesupport for Loan Equipment project [10] compared in-person support for computer access teams with remote support. The ACE runs a technology loan library and has observed a need for “good quality, on-going training while a piece of equipment is borrowed [10].” Results indicated that telerehabilitation interventions provided the following benefits to users and aides [10]:

  • clients could interact in real-time with clinicians;

  • trained clinicians could configure devices remotely;

  • training and support via videoconferencing technology was perceived to be of higher quality than more conventional communication methods (e.g., phone, email).

Also in Europe, the Remote Service of Rehabilitation Technology (RESORT) project [7779] developed telerehabilitation technology that allows a clinician to establish a communication channel with a client for transmitting audio and visual data and for desktop sharing1. A clinician can use RESORT to interact with a client, monitor (but not collect data on) the client’s computer use, and adjust the configuration of the software. The RESORT team also developed an application programming interface (API) that allows an application on the client’s computer to be synchronized with the clinician’s computer. This API is different from other application sharing technologies (e.g., Microsoft NetMeeting) in that it supports real-time synchronization.

4 DISCUSSION

Quantitative evaluation tools such as Compass provide the means for clinicians to collect evidence in a rigorous manner. The underlying technology, however, can be advanced in several directions. One opportunity for potential improvement is the integration of additional performance measures, many of which were developed by the human-computer interaction (HCI) community. For example, MacKenzie and colleagues [80] have developed error measures that compare the cursor’s actual path of travel to the “ideal” path of travel. Keates and Hwang [8185] extended this work by developing error measures that incorporated the “instantaneous” optimal cursor path, referred to as the Instantaneous Task Axis (ITA). These measures can provide the client and clinician with a great deal of additional information, but it is likely that a distinct subset of measures will be of greatest interest for each client, device and task. Hence, a means of flexibly integrating user- and device-specific errors would also be useful.

Clinical assessment tools can also benefit by adopting protocols originally developed for HCI research. For example, investigators within the field of human factors have recently begun employing “unconstrained” text entry protocols [8690] in which the user is allowed to make errors and to decide whether or not to correct the errors that occur. The primary advantage of this approach is that it allows users to enter text under more natural, realistic conditions [8690]. This approach also allows investigators to analyze the entire input stream, including errors and error corrections [90], thus providing a more detailed picture of text entry.

Wobbrock has recently demonstrated how the text input stream can be decomposed into [90]:

  • Non-errors. Correct keystrokes (i.e., the correct key pressed at the correct time).

  • Substitutions. Incorrect keystrokes in which one character is entered instead of another.

  • Insertions. Incorrect keystrokes in which additional characters are entered.

  • Omissions. Errors in which a character that should appear in the input stream does not.

  • Fixes. Keystrokes used to remove characters or reposition the cursor (e.g., backspace, delete, arrow keys)

The algorithms developed by Wobbrock, however, make certain assumptions that are not necessarily valid for individuals with disabilities or for text entry methods like word completion and abbreviation expansion where a single input can generate multiple characters [90]. Additional work is needed to determine how Wobbrock’s work can be extended to cover these situations.

Software like IDA may one day assist clinicians in configuring CAT, and may also be used by clients to adjust the configuration of their technology over time. IDA is limited, however, in that the client must complete performance tests in order for IDA to collect enough data to make configuration recommendations. A more desirable approach would be for IDA to make its recommendations based on the client’s performance on unconstrained real-world tasks in the normal course of computer use. Hurst and colleagues have used machine learning techniques to automatically categorize user performance on pointing tasks based on whether or not assistance with clicking on targets was likely to improve performance [91].

Table 2.

Keyboard configuration parameters.

Parameter Description
Repeat Delay How long a key must be held down before it begins to repeat.
Repeat Rate Once the keyboard begins to repeat a character, the rate at which it repeats.
SlowKeys How long a key must be held down before it is accepted.
BounceKeys Tells the operating system to ignore keystrokes that are depressed within x seconds of the previous key release.
StickyKeys When StickyKeys are activated, the typist can enter key combinations (e.g., Shift-A to type a capital A) by pressing the modifier key (e.g., Shift) and other keys (e.g., “A”) in series, rather than holding down multiple keys simultaneously.
ToggleKeys Gives an auditory signal when locking keys, such as Caps Lock, are depressed.

ACKNOWLEDGEMENTS

Funding provided by the National Science Foundation (#0540865) and the National Institutes of Health (#2R44HD045015-03)

This work was supported by the National Science Foundation (#0540865) and the National Institutes of Health (#2R44HD045015-03)

Footnotes

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

1

Desktop sharing refers to the ability of one computer user to remotely operate another computer while seeing the windows and icons on the remote computer’s “desktop”

REFERENCES CITED

  • 1.Anson DK. Alternative Computer Access: A Guide to Selection. Philadelphia, PA: F.A. Davis Company; 1997. [Google Scholar]
  • 2.Chen C-L, Chen H-C, Cheng P-T, Chen C-Y, Chen H-C, Chou S-W. Enhancement of Operational Efficiencies for People With High Cervical Spinal Cord Injuries Using a Flexible Integrated Pointing Device Apparatus. Archives of Physical Medicine and Rehabilitation. 2006;87:866–873. doi: 10.1016/j.apmr.2006.02.025. [DOI] [PubMed] [Google Scholar]
  • 3.Dobransky K, Hargittai E. The disability divide in Internet access and use Information. Communication & Society. 2006;9:313–334. [Google Scholar]
  • 4.Drainoni M, Houlihan B, Williams S, Vedrani M, Esch D, Lee-Hood E, Weiner C. Patterns of Internet use by persons with spinal cord injuries and relationship to health-related quality of life. Archives of Physical Medicine and Rehabilitation. 2004;85:1872–1879. doi: 10.1016/j.apmr.2004.07.350. [DOI] [PubMed] [Google Scholar]
  • 5.Madara E. The mutual-aid self-help online revolution. Social Policy. 1997;27:20–26. [Google Scholar]
  • 6.McKenna K, Seidman G. In: The social net: Human behavior in cyberspace. Amichai-Hamburger Y, editor. Oxford, UK: Oxford University Press; 2005. pp. 191–217. [Google Scholar]
  • 7.McKenna K, Bargh J. Plan 9 from cyberspace: The implications of the Internet for personality and social psychology. Personality and Social Psychology Review. 2000;4:57. [Google Scholar]
  • 8.Morahan-Martin J, Schumacher P. Loneliness and social use of the Internet. Computers in Human Behavior. 2003;19:656–671. [Google Scholar]
  • 9.Stevenson B, McQuivey JL. Cambridge, MA: A Research Study Commissioned by Microsoft Corporation and Conducted by Forrester Research; 2003. The Wide Range of Abilities and Its Impact on Computer Technology. [Google Scholar]
  • 10.Hazell G, Colven D. Oxford: ACE Centre Advisory Trust; 2001. ACE Centre Telesupport for Loan Equipment; p. 35. [Google Scholar]
  • 11.Raskind M. Assistive technology and adults with learning disabilities: A blueprint for exploration and advancement. Learning Disability Quarterly. 1993;16:185–196. [Google Scholar]
  • 12.Scherer MJ. What we know about women's technology use, avoidance, and abandonment. Women and Therapy. 1993;14:117–129. [Google Scholar]
  • 13.Scherer MJ, Galvin JC. In: Evaluating, selecting and using appropriate assistive technology. Galvin JC, Scherer MJ, editors. Gaithersburg, MD: Aspen Publication; 1996. pp. 1–26. [Google Scholar]
  • 14.Phillips B, Zhao H. Predictors of Assistive Technology Abandonment. Assistive Technology. 1993;5:36–45. doi: 10.1080/10400435.1993.10132205. [DOI] [PubMed] [Google Scholar]
  • 15.Tewey BP, Barnicle K, Perr A. The wrong stuff. Mainstream. 1994;19:19–23. [Google Scholar]
  • 16.Lysley A, Colven D, Donegan M. Oxford, England: ACE Centre; 1999. CATCHNET Final Report. [Google Scholar]
  • 17.Batavia AJ, Dillard D, Phillips B. Washington, D.C: National Institute on Disability and Rehabilitation Research; 1990. How to avoid technology abandonment. [Google Scholar]
  • 18.Phillips B, Zhao H. Predictors of assistive technology abandonment. Assistive Technology. 1993;5:36. doi: 10.1080/10400435.1993.10132205. [DOI] [PubMed] [Google Scholar]
  • 19.Riemer-Reiss M, Wacker R. Assistive Technology Use and Abandonment among College Students with Disabilities`. International Electronic Journal for Leadership in Learning. 1999;3:23. [Google Scholar]
  • 20.Schartz K, Schartz H, Blanck P. Employment of persons with disabilities in information technology jobs: literature review for "IT works". Behav. Sci. Law. 2002;20:637–657. doi: 10.1002/bsl.510. [DOI] [PubMed] [Google Scholar]
  • 21.Stevenson B, McQuivey JL. Cambridge, MA: A Research Study Commissioned by Microsoft Corporation and Conducted by Forrester Research; 2003. Examining Awareness, Use, and Future Potential. [Google Scholar]
  • 22.Trewin S, Pain H. Keyboard and mouse errors due to motor disabilities. International Journal of Human-Computer Studies. 1999;50:109–144. [Google Scholar]
  • 23.Riemer-Reiss M. Factors Associated with Assistive Technology Discontinuance Among Individuals with Disabilities. Journal of Rehabilitation. 2000 [Google Scholar]
  • 24.Scherer M, Cushman L. Measuring subjective quality of life following spinal cord injury: A validation study of the assistive technology device predisposition assessment. Disability and Rehabilitation. 2001;23:387–393. doi: 10.1080/09638280010006665. [DOI] [PubMed] [Google Scholar]
  • 25.Barker M, Cook AM. A systematic approach to evaluating physical ability for control of assistive devices; International Conference on Assistive Technology for People with Disabilities; RESNA: RESNA Press; 1981. pp. 287–289. [Google Scholar]
  • 26.Rosen M, Goodenough-Trepagner C. The Tufts-MIT prescription guide: Assessment of users to predict the suitability of augmentative communication devices. Assistive Technology. 1989;1:51–61. [Google Scholar]
  • 27.Bowser G, Zabala J. ConnSENSE Bulletin. 2005. SETT & Re-SETT: Concepts of AT implementation. [Google Scholar]
  • 28.Cushman LA, Scherer MJ. Measuring the relationship of assistive technology use, functional status over time, and consumer-therapist perceptions of ATs. Assistive Technology. 1996;8:103–109. doi: 10.1080/10400435.1996.10132280. [DOI] [PubMed] [Google Scholar]
  • 29.Scherer MJ, Cushman LA. Predicting satisfaction with assistive technology for a sample of adults with new spinal cord injuries. Psychology Report. 2000;87:981–987. doi: 10.2466/pr0.2000.87.3.981. [DOI] [PubMed] [Google Scholar]
  • 30.Scherer MJ, Cushman LA. Measuring subjective quality of life following spinal cord injury: a validation study of the assistive technology device predisposition assessment. Disability and Rehabilitation. 2001;23:387–393. doi: 10.1080/09638280010006665. [DOI] [PubMed] [Google Scholar]
  • 31.Chambers AC. CASE/TAM assistive technology policy and practice series: Has technology been considered?; Annual Meeting of the Technology and Media Division of the Council for Exceptional Children; 1997; Reston, VA: [Google Scholar]
  • 32.Reed P, editor. Assessing Students' Needs for Assistive Technology. Oshkosh, WI: Wisconsin Assistive Technology Initiative; 2000. [Google Scholar]
  • 33.Reed P, Bowser G. Education Tech Points: A framework for assistive technology planning and systems change in schools; Conference on Technology and Persons with Disabilities (CSUN); 1998; Los Angeles, CA: CSUN; [Google Scholar]
  • 34.Anson D. Alternative Computer Access: A Guide to Selection. Philadelphia, PA: F.A. Davis Company; 1997. [Google Scholar]
  • 35.Lee KS, Thomas DJ. Control of Computer-Based Technology for People with Physical Disabilities: An Assessment Manual. Toronto: University of Toronto Press; 1990. [Google Scholar]
  • 36.Dumont C, Vincent C, Mazer B. Development of a standardized instrument to assess computer access performance. American Journal of Occupational Therapy. 2002;56:60–68. doi: 10.5014/ajot.56.1.60. [DOI] [PubMed] [Google Scholar]
  • 37.Williams WB, Stemach G, Stanger C. Lifespace Access Profile: Assistive Technology Assessment and Planning for Individuals with Severe or Multiple Disabilities. Irvine, CA: Lifespace Access Assistive Technology Systems; 1995. [Google Scholar]
  • 38.Stapleton D, Garrett R, Seeger B. VOCASelect and Computer Access Selector: Software tools to assist in choosing assistive technology; International Conference on Assistive Technology for People with Disabilities (RESNA); 1997; Pittsburgh, PA: RESNA Press; [Google Scholar]
  • 39.Lahm EA, Gassaway LJ. Matching assistive technology to the individual using an expert system; Annual Meeting of the Technology and Media Division of the Council for Exceptional Children; 2003; Reston, VA: [Google Scholar]
  • 40.Lahm EA, Gassaway LJ, Reed AG. ATES: Assistive Technology Expert System; Annual Meeting of the Technology and Media Division of the Council for Exceptional Children; 2002; Washington, DC: [Google Scholar]
  • 41.LoPresti EF, Koester HH, McMillan W, Moore P, Ashlock G, Simpson RC. Compass: Software for Computer Skills Assessment; International ACM Conference on Assistive Technologies (Assets); 2002; Edinburgh, Scotland:: ACM; [Google Scholar]
  • 42.Koester HH, McMillan W. Usability testing of software for assessing computer usage skills; International Conference on Assistive Technology for People with Disabilities; RESNA: RESNA Press; 1998. [Google Scholar]
  • 43.Koester HH, McMillan W. Software for assessing computer usage skills; International Conference on Assistive Technology for People with Disabilities; 1997; RESNA: RESNA; [Google Scholar]
  • 44.Ashlock G, Koester HH, LoPresti EF, McMillan WW, Simpson RC. User-centered design of software for assessing computer usage skills. In: Simpson RC, editor. International Conference on Assistive Technology for People with Disabilities (RESNA); 2003; Atlanta, GA: RESNA; [Google Scholar]
  • 45.Trewin S, Pain H. A model of keyboard configuration requirements. Behaviour & Information Technology. 1999;18:27–35. [Google Scholar]
  • 46.Lesher GW, Higginbotham DJ, Moulton BJ. Techniques for automatically updating scanning delays; RESNA 2000 Annual Conference; Orlando, FL: RESNA Press; 2000. pp. 75–77. [Google Scholar]
  • 47.LoPresti EF, Brienza D. Adaptive Software for Head-Operated Computer Controls. IEEE Transactions on Neural Systems and Rehabilitation Engineering. 2004;12:102–111. doi: 10.1109/TNSRE.2003.822762. [DOI] [PubMed] [Google Scholar]
  • 48.McGill R, O'Beirne H, Milner M, Naumann S. Towards the development of adaptive user interfaces for tremor disabled persons: A blackboard expert system approach; Annual International Conference of the IEEE Engineering in Medicine and Biology Society; 1991; IEEE Press; pp. 1262–1263. [Google Scholar]
  • 49.Simpson RC, Koester HH. Adaptive One-Switch Row-Column Scanning. IEEE Transactions on Rehabilitation Engineering. 1999;7:464–473. doi: 10.1109/86.808950. [DOI] [PubMed] [Google Scholar]
  • 50.Tracey M, Winters J. Neuro-Fuzzy Advisor for Mouse Setting in Microsoft Windows; Proceedings of the First Joint BMES/EMBS Conference; 1999; Atlanta, GA: IEEE; p. 664. [Google Scholar]
  • 51.Trewin S. Automating accessibility: The dynamic keyboard; Assets ‘04: Proceedings of the 6th International ACM SIGACCESS Conference on Computers and Accessibility; Atlanta, GA: ACM; 2004. pp. 71–78. [Google Scholar]
  • 52.Trewin S. Extending keyboard adaptability: An investigation. Universal Access in the Information Society. 2002;2:44–55. [Google Scholar]
  • 53.Trewin S. Configuration agents, control and privacy; Conference on Universal Usability 2000; 2000; New York, NY: ACM; pp. 9–16. [Google Scholar]
  • 54.Trewin S. ASSETS 2002. New York, NY: ACM; 2002. An invisible keyguard; pp. 143–149. [Google Scholar]
  • 55.Trewin S. ASSETS 2004. New York, NY: ACM; 2004. Automating accessibility: The dynamic keyguard; pp. 71–78. [Google Scholar]
  • 56.Trewin S, Pain H. Dynamic modeling of keyboard skills: Supporting users with motor disabilities; 6th International Conference on User Modeling; 1997; Wien, NY: Springer; pp. 135–146. [Google Scholar]
  • 57.Trewin S, Pain H. A model of keyboard configuration requirements. Behavior and Information Technology. 1999;18:27–35. [Google Scholar]
  • 58.Trewin S, Pain H. Dynamic Modelling of Keyboard Skills: Supporting Users With Motor Disabilities; Sixth International Conference on User Modeling; 1997. pp. 135–146. [Google Scholar]
  • 59.Koester HH, LoPresti EF, Simpson RC. Toward automatic adjustment of keyboard settings for people with physical impairments. Disability and Rehabilitation: Assistive Technology. 2007;2:261–274. doi: 10.1080/17483100701284265. [DOI] [PubMed] [Google Scholar]
  • 60.LoPresti EF, Koester HH, Simpson RC. Toward automatic adjustment of pointing device configuration to accommodate physical impairment. Disability and Rehabilitation: Assistive Technology. 2008;3:221–235. doi: 10.1080/17483100801972975. [DOI] [PubMed] [Google Scholar]
  • 61.Damper RI. Text composition by the physically disabled: A rate prediction model for scanning input. Applied Ergonomics. 1984;15:289–296. doi: 10.1016/0003-6870(84)90201-1. [DOI] [PubMed] [Google Scholar]
  • 62.Koester HH, Levine SP. Modeling the speed of text entry with a word prediction interface. IEEE Transactions on Neural Systems and Rehabilitation Engineering. 1994;2:177–187. [Google Scholar]
  • 63.Cronk SR, Schubert RW. Development of a Real-Time Expert System for Automatic Adaptation of Scanning Rates; Annual Conference on Rehabilitation Technology (RESNA); 1987; San Jose, CA: RESNA Press; [Google Scholar]
  • 64.Lesher GW, Higginbotham J, Moulton BJ. Techniques for automatically updating scanning delays; Annual Conference on Rehabilitation Technology (RESNA); 2000; Orlando, FL: RESNA Press; [Google Scholar]
  • 65.Lesher GW, Moulton BJ, Higginbotham J, Alsofrom B. Acquisition of scanning skills: The use of an adaptive scanning delay algorithm across four scanning displays; Annual Conference on Rehabilitation Technology (RESNA); 2002; Minneapolis, MN: RESNA Press; [Google Scholar]
  • 66.Simpson RC, Koester HH. Adaptive one-switch row-column scanning. IEEE Transactions on Neural Systems and Rehabilitation Engineering. 1999;7:464–473. doi: 10.1109/86.808950. [DOI] [PubMed] [Google Scholar]
  • 67.Simpson RC, Koester HH, LoPresti EF. Evaluation of an adaptive row/column scanning system. Technology and Disability. 2007;18:127–138. [Google Scholar]
  • 68.Field MJ. Telemedicine: A Guide to Assessing Telecommunications in Health Care. Washington, D.C: National Acadamy Press; 1996. [PubMed] [Google Scholar]
  • 69.Burns RB, Crislip D, Daviou P, Temkin A, Vesmarovich S, Anshutz J, Furbish C, Jones ML. Using telerehabilitation to support assistive technology. Assist Technol. 1998;10:126–133. doi: 10.1080/10400435.1998.10131970. [DOI] [PubMed] [Google Scholar]
  • 70.Magaziner J, Zimmerman SI, Gruber-Baldini AL, Hebel JR, Fox KM. Proxy reporting in five areas of functional status: Comparison with self-reports and observations of performance. American Journal of Epidemiology. 1997;146:418–428. doi: 10.1093/oxfordjournals.aje.a009295. [DOI] [PubMed] [Google Scholar]
  • 71.Rubenstein LZ, Schairer C, Wieland GD, Kane R. Systematic biases in functional status assessment of elderly adults: Effects of different data sources. Journal of Gerentology. 1984;39:686–691. doi: 10.1093/geronj/39.6.686. [DOI] [PubMed] [Google Scholar]
  • 72.Sager MA, Dunham NC, Schwantes A, Mecum L, Halverson K, Harlowe D. Measurement of activities of daily living in hospitalized elderly: A comparison of self-report and performance-based methods. Journal of American Geriatrics. 1992;40:457–462. doi: 10.1111/j.1532-5415.1992.tb02011.x. [DOI] [PubMed] [Google Scholar]
  • 73.Cooper RA, Fitzgerald SG, Boninger ML, Brienza DM, Shapcott N, Cooper R, Flood K. Telerehabilitation: Expanding access to rehabilitation expertise; Proceedings of the IEEE; 2001; pp. 1174–1191. [Google Scholar]
  • 74.Burns R, Hauber R, Vesmarovich S. Telerehabilitation: continuing cases and new applications. In: Winters J, editor. The RESNA 2000 Annual Conference. Technology for the New Millennium; 2000; Orlando, FL: pp. 258–260. [Google Scholar]
  • 75.Cole E, Ziegmann M, Wu Y, Yonker V, Gustafson C, Cirwithen S. User of ‘Therapist-Friendly’ Tools in Cognitive Assistive Technology and Telerehabilitation; The RESNA International Conference; 2000; Orlando, FL: RESNA; pp. 31–33. [Google Scholar]
  • 76.Donegan M. Oxford, England: ACE Centre; 2002. The TELENET Project Summary Final Report. [Google Scholar]
  • 77.Panek P, Zagler WL, Beck C, Hine N, Seisenbacher G, Stefankis N. Vienna International Workshop on Distance Education and Training. Vienna, Austria: 2001. Providing Tele-Support and Tele-Training to Severely Disabled Persons Using IP-Based Networks. [Google Scholar]
  • 78.Panek P, Zagler WL. Vienna, Austria: Fortec, Vienna University of Technology; 2001. Remote Service of Rehabilitation Technology Final Report. [Google Scholar]
  • 79.Panek P, Beck C, Hochgatterer A, Mina S, Prazak B, Seisenbacher G, Soede M, Zagler WL. Tele-help and remote service provision using RESORT prototype system. In: Miesenberger K, et al., editors. 8th International Conference on Computers Helping People with Special Needs; 2002; Austria: Linz; pp. 635–642. [Google Scholar]
  • 80.MacKenzie IS, Kauppinen T, Silfverberg M. Accuracy measures for evaluating computer pointing devices; SIGCHI Conference on Human Factors in Computing Systems; 2001; Seattle, WA: ACM Press; [Google Scholar]
  • 81.Hwang F. A study of cursor trajectories of motion-impaired users. In: Terveen L, Wixon D, editors. CHI '02 Extended Abstracts on Human Factors in Computing Systems. Minneapolis, MN: ACM Press; 2002. pp. 842–843. [Google Scholar]
  • 82.Hwang F. Partitioning cursor movements in "point and click" tasks. In: Cockton G, Korhonen P, editors. CHI '03 Extended Abstracts on Human Factors in Computing Systems. Ft. Lauderdale, FL: ACM Press; 2003. pp. 682–683. [Google Scholar]
  • 83.Hwang F, Keates S, Langdon P, Clarkson J. Mouse movements of motion-impaired users: A submovement analysis; 6th International ACM SIGACCESS Conference on Computers and Accessibility; 2004; Atlanta, GA: ACM Press; [Google Scholar]
  • 84.Keates S, Hwang F, Langdon P, Clarkson PJ, Robinson P. Cursor measures for motion-impaired computer users. In: Hanson VL, J Jacko A, editors. Proceedings of the Fifth International ACM Conference on Assistive Technologies; 2002; Edinburgh, Scotland: ACM Press; pp. 135–142. [Google Scholar]
  • 85.Keates S, Hwang F, Langdon P, Clarkson PJ, Robinson P. The user of cursor measures for motion-impaired computer users. Universal Access in the Information Society. 2002;2:18–29. [Google Scholar]
  • 86.MacKenzie IS, Soukoreff RW. A character-level error analysis technique for evaluating text entry methods. In: Bertelsen OW, editor. Proceedings of the Second Nordic Conference on Human-Computer Interaction; 2002; Aarhus, Denmark: ACM Press; pp. 243–246. [Google Scholar]
  • 87.Soukoreff RW, MacKenzie IS. Measuring errors in text entry tasks: An application of the Levenshtein string distance statistic. In: Tremaine MM, editor. CHI '01 extended abstracts on Human factors in computing systems. Seattle, WA: ACM Press; 2001. pp. 319–320. [Google Scholar]
  • 88.Soukoreff RW, MacKenzie IS. Metrics for text entry research: An evaluation of MSD and KSPC, and a new unified error metric. In: Cockton G, Korhonen P, editors. Proceedings of the SIGCHI Conference on Human Factors in Computing Systems; 2003; Ft. Lauderdale, FL: ACM Press; pp. 113–120. [Google Scholar]
  • 89.Soukoreff RW, MacKenzie IS. Recent developments in text-entry error rate measurement. In: Dykstra-Erickson E, Tscheligi M, editors. CHI '04 Extended Abstracts on Human Factors in Computing Systems. Vienna, Austria: ACM Press; 2004. pp. 1425–1428. [Google Scholar]
  • 90.Wobbrock JO, Myers BA. Analyzing the input stream for character-level errors in unconstrained text entry evaluations. ACM Transactions on Computer-Human Interaction. 2006;13:458–489. [Google Scholar]
  • 91.Hurst A, Hudson SE, Mankoff J, Trewin S. Automatically detecting pointing performance; 13th International Conference on Intelligent User Interfaces; Gran Canaria, Spain: ACM; 2008. pp. 11–19. [Google Scholar]

RESOURCES