Abstract
For more than a century, measurement experts have distinguished between time-limited tests and untimed power tests, which are administered without time limits or with time limits so generous that all students are assured of completing all items. On untimed power tests, students can differ in their propensity to correctly respond to every item, and items should differ in how many correct responses they elicit. However, differences among students’ speed of responding do not confound untimed power tests; therefore, untimed power tests ensure more accurate assessment. In this article, we present four empirically based reasons to administer untimed power tests rather than time-limited tests in educational settings. (1) Time-limited tests are less valid; students’ test-taking pace is not a valid reflection of their knowledge and mastery. (2) Time-limited tests are less reliable; estimates of time-limited tests’ reliability are artificially inflated due to artifactual consistency in students’ rate of work rather than authentic consistency in students’ level of knowledge. (3) Time-limited tests are less inclusive; time-limited tests exclude students with documented disabilities who, because they are legally allowed additional test-taking time, are often literally excluded from test-taking classrooms. (4) Time-limited tests are less equitable; in addition to excluding students with documented disabilities, time-limited tests can also impede students who are learning English, students from underrepresented backgrounds, students who are older than average, and students with disabilities who encounter barriers (e.g., stigma and financial expense) in obtaining disability documentation and legally mandated accommodations. We conclude by offering recommendations for avoiding time-limited testing in higher educational assessment.
Keywords: disability, equity, higher education, inclusion, timed testing
“Yeah, everywhere around the country. What happened is, all the wealthy families … figured out that if I get my kid tested and they get extended time, they can do better on the test”
[Cooperating Witness-1, from United States Department of Justice, 2019, p. 25]
For more than a century, measurement experts have distinguished between time-limited tests and untimed power tests (Lorge, 1936; McCall, 1916; Ruch & Koerth, 1923).1 Power tests are typically administered without any time limits or with time limits so generous that all students are assured of having plenty of time to complete every item (Sisk, 1926). Students taking an untimed power test will most likely differ in their propensity (or preparation) to correctly respond to every item, and items on an untimed power test should be designed to differ in how many correct responses they elicit (Kyllonen & Zu, 2016). However, differences among students’ speed of responding do not confound untimed power tests; therefore, untimed power tests ensure more accurate assessment (Colker, 2019; Donlon, 1975).
Untimed power tests have been regarded, for nearly a century, as “the ideal” (Peak & Boring, 1926, p. 72) because they best assess the “complexity of … thought-processes” (Line & Kaplan, 1932, p. 1). In contrast, time-limited tests are good only for assessing tasks that are so cognitively simple that everyone would respond correctly if no time limit was imposed (Danthiir, Wilhelm, & Schacht, 2005; DuBois, 1932; Mead & Drasgow, 1993).
For example, on the Differential Aptitude Tests only lower-level clerical speed and accuracy are measured with time limits; higher-level abstract reasoning, verbal analysis, spatial relations, and mechanical thought are assessed via untimed power tests (Bennett, Seashore, & Wesman, 1956). In this article, we present four empirically based reasons not to administer time-limited tests in higher educational settings and to instead administer untimed power tests. Our reasons are that time-limited tests are less valid, less reliable, less inclusive, and less equitable.
Time-Limited Tests Are Less Valid
For nearly a century, we have known that students’ pace on an untimed power test does not validly reflect their performance. Some students work quickly but perform poorly; other students work slowly but perform well, and all variations exist in between (Mudge, 1921). The lack of prediction from pace to performance has been documented across a variety of untimed power tests and a wide range of large student samples, including grade-school and high-school students (Dodonova & Dodonov, 2013; Freeman, 1923; Line & Kaplan, 1932; Mangan, 1959; Slater, 1938; Terranova, 1972; Vernon & Kantor, 1986), as well as college students (Dodonova & Dodonov, 2013; DuBois, 1932; Freeman, 1923, 1932; Lord, 1956; Miller & Weiss, 1976; Neubauer, 1990; Tinker, 1944; Tryon & Jones, 1933; White, 1973; Yates, 1963, 1966).
In the classroom, one can observe the poor prediction of test-taking pace to test-taking performance by simply correlating the order in which students finish an exam with the grade they earn on that exam. Indeed, nearly a century ago, two psychology professors did just that. Longstaff and Porter (1928, p. 638) described their motivation in the following way:
We hear school men very authoritatively saying that the fast students make the best grades and the slow ones the poorest. … Statements of this kind are usually based on the assumption that if a student knows the subject in which he is being tested it should follow that he requires but a short time to make his answer. Needless to say, this assumption merits confirmation.
Yet, confirmation was not obtained. Rather, Longstaff and Porter’s (1928) data, based on nearly 200 General Psychology students each taking nearly 15 exams, demonstrated “very little, if any, relationship between the times required and the scores made” (p. 641).
Forty years later, Burack (1967, p. 164) reminisced that “psychology majors in the 1930s” had read studies demonstrating the poor prediction of test-taking pace to test-taking performance in psychology courses and wondered whether the prediction still failed; it did (see also Michael & Michael, 1969). Across the next 30 years, the prediction continued to fail in psychology courses across the country (Becker & Suls, 1982; Bridges, 1985; Ebel & Frisbie, 1991; Foos, 1989; Herman, 1997; Johnston, 1987; Kennedy, 1994; Lester, 1991; Paul & Rosenkoetter, 1980; Wierzbicki, 1994).
Into the 21st century, the prediction continues to fail not only in psychology courses (Brothen, 2012; Hammonds & Mariano, 2015) but also in chemistry (Ghaffari, 2015; Nevo & Spector, 1979); engineering (Armitage, 1999); agriculture (McDannell & Lyvers Peffer, 2013); business, including finance, real estate, and management (Beaulieu & Frost, 1994; Lee, 2012; Schnusenberg & Slater, 2011); English and literature (Lovett, Lewandowski, & Potts, 2017; Rogers, 1968); statistics (Hammonds & Mariano, 2015; Onwuegbuzie & Daley, 1996); and mathematics (Rogers, 1968). Why does the prediction fail? The prediction fails because putting time limits on power tests introduces irrelevant variance (Ackerman & Ellingsen, 2016; Donlon, 1975; Lu & Sireci, 2007; Peterson, 1993; Pitoniak & Royer, 2001; Thissen, 1983).
The empirical evidence against the assumption that “fast students make the best grades” comes from thousands of students taking all types of tests: graduate as well as undergraduate examinations (Freeman, 1923; Michael & Michael, 1969; Onwuegbuzie & Daley, 1996); open-book as well as closed-book exams (Mi-chael & Michael, 1969); and exams comprising problem sets (Armitage, 1999), essays (Burack, 1967; Foos, 1989; Lee, 2012), short-answer (Burack, 1967; Onwuegbuzie & Daley, 1996; Wierzbicki, 1994), as well as multiple-choice items (Burack, 1967; Bridges, 1985; Foos, 1989; Freeman, 1923; Longstaff & Porter, 1928; Paul & Rosenkoetter, 1980; Schnusenberg & Slater, 2011; Wierzbicki, 1994).
Why does this erroneous assumption persist?2 It persists from faulty reasoning and misaligned analogies. For example, a recent OpEd likened professors providing extended time on classroom exams, which is a topic we address later, to World Track and Field officials providing Usain Bolt’s competitor “a 20-meter head start” (Pardy, 2017). But exams are not track meets, and we have known for more than a century, as Whipple documented in his classic Manual of Mental and Physical Tests (Whipple, 1914, p. 10), that “if we seek to evaluate the complex ‘higher’ mental functions, speed is not the primary index of efficiency, as is borne out by the evidence that speed and intelligence are not very highly correlated.”
We have also known for nearly a century that beyond classrooms and into occupations, skill also uncouples from pace. For example, Lemmon (1927) reported a poor correlation between trained pilots’ response speed and their aviation ability, akin to the poor correlation between students’ test-taking pace and their test-taking performance. Berliner (1994) identified professional tasks on which novice teachers required more time but other tasks on which experienced teachers required more time. Similarly, because more knowledgeable students do not take less time on time-limited tests and time-limited tests become less valid as they tap more intellectual content (Whipple, 1914), we recommend untimed power tests in higher education settings.
Time-Limited Tests Are Less Reliable
For almost as long as measurement experts have distinguished time-limited tests from untimed power tests, experts have cautioned about the spurious reliability of time-limited tests (Angoff, 1953; Cooper, 1984; Cronbach & Warrington, 1951; Thissen, 1983; Wesman, 1949). For example, in their classic textbook, Introduction to Classical and Modern Test Theory, Crocker and Algina (1986) warned that reliability estimates of time-limited tests are artificially inflated due to artifactual consistency in students’ rate of work rather than authentic consistency in students’ level of knowledge. Thus, just as variation between students’ time-limited performance confounds estimates of time-limited tests’ validity, consistency within students’ time-limited performance inflates estimates of time-limited tests’ reliability (Cronbach, 1949).
Anne Anastasi, psychology’s “test guru” (Goode, 2001), issued more dire warnings, noting “It has been repeatedly demonstrated, both theoretically and empirically,” that the reliability coefficients of time-limited tests “may be completely meaningless” when computed using standard procedures (Anastasi & Drake, 1954, p. 529). Although some statistical corrections for computing the reliability of time-limited tests have been suggested, we have known for decades that the best way to improve a time-limited test’s reliability is simply to remove its time limits (Carter, 1932; Traub & Hambleton, 1972).
We suppose most instructors would like to administer untimed tests. We assume most instructors want to assess students’ level of acquired knowledge (DuBois, 1932; Tinker, 1944) not their speed in reacting to test items (Mead & Drasgow, 1993; Peak & Boring, 1926). We believe most instructors prize genuinely reliable untimed power tests over artificially reliable time-limited tests, although some instructors continue to erroneously conflate speed with power. Yet, we observe many instructors inadvertently implementing what should be untimed power tests as less genuinely reliable time-limited tests solely for administrative reasons (Donlon, 1975, 1984; Kerstiens, 1990; Lovett, 2010; Lu & Sireci, 2007; Mollenkopf, 1950; National Association of Secondary-School Principals, 1948; Parr, Levi, & Jacka, 1996; Pitoniak & Royer, 2001; Wesman, 1949).
As Morrison (1960, p. 231) wrote decades ago, “the popularity of time-limit tests is due more to their practical administrative advantages … than to any experimentally-supported rationale governing the imposition of time limits on performance.” Class periods are finite; even the typically more generous final exam periods are bound. These time limits introduce a “spurious … contribution” (Wilhelm & Schulze, 2002, p. 537) that can “skew measures” (Henderson, 2005) to produce “severe problems” (Peterson, 1993, p. 9) and alter the “construct the test intends to measure” (Lu & Sireci, 2007, p. 31). Therefore, we reiterate Mollenkopf’s (1960, p. 223) decades-old admonition: “you should have a better reason for setting whatever [time] limits you impose than mere convenience of administration.”
Time-Limited Tests Are Less Inclusive
The faulty assumption that “fast students make the best grades and the slow ones the poorest” has also disparaged expectations about students with disabilities. For decades, slow was a euphemism for intellectual disability (Abrams & Goodman, 1998). The late twentieth-century term mental retardation, which referred to what we now call intellectual disability, meant “an abnormal slowness of thought or action” (Merriam-Webster, n.d.) and was derived from the Latin term retardāre (“to make slow,” The American Heritage Dictionary, n.d.). An early, but mistaken, view of intelligence promoted this erroneous assumption: “To be quick is to be intelligent, while slowness is synonymous with dullness” (Line & Kaplan, 1932, p. 1).
Even in the 21st century this unfortunate assumption persists, as demonstrated by the 2019 college entrance scandal, during which affluent parents illegally obtained college admission for their children and from which we quoted at the outset of our article. One parent was instructed by the alleged fixer “to tell [their daughter] when she gets tested … to be stupid, not to be as smart as she is. The goal is to be slow, to be not as bright” (United States Department of Justice, 2019, p. 25, emphasis added).
One of the loopholes exploited in the college entrance scandal capitalized on U.S. laws that protect the rights of students with disabilities, including the Rehabilitation Act of 1973, in particular, Section 504, and the Americans with Disabilities Act of 1990, in particular, Section 309, which extended disability anti-discrimination protections to the private sector (United States Government Accountability Office, 2012). Simply put: Students with disabilities must be provided reasonable testing accommodations, including Braille, sign language interpretation, large print, auditory presentation and description, and extended time on time-limited tests (Casey, 1987; Centra, 1986; Elliott, McKevitt, & Kettler, 2002; Hishinuma, 1998; Phillips, 1994; Pitoniak & Royer, 2001; Thurlow, Seyfarth, Scott, & Ysseldyke, 1997; United States Department of Justice, 2014).
Extended time on time-limited tests is typically operationalized as “time and a half,” meaning that students provided this accommodation are allowed 50% more time to complete their tests (Laitusis, Morgan, Bridgeman, Zanna, & Stone, 2007; United States Government Accountability Office, 2012). However, most students who receive extended-time accommodations do not use all the additional 50% testing time.
For example, in one study of 825 students with learning disabilities at a small arts and technical college, very few students who were provided extended time used more than the standard time, and those who did rarely used more than an additional 25% (Holmes & Silvestri, 2019). Similarly, in another study of 605 students with disabilities at a regional university, most students who were provided extended time used only the standard time, students with psychiatric disabilities using more (around half of the additional 50%; Spenceley & Wheeler, 2016). In another study of 50 students with a variety of learning disabilities taking the SAT, no student used more than 14% extra time (Cahalan-Laitusis, King, Cline, & Bridgeman, 2006).
As Gernsbacher (2015, p. 35) wrote, “When students request extended time or time and a half, what they are really requesting is not to feel the pressure of time ticking off; not to experience anxiety about running out of time; not to have [an untimed] power test administered as a [time-limited] test” (see also Hadley, 2007). The studies demonstrating how little additional time students with disabilities actually use support this interpretation. In essence, students with disabilities who are provided extended time are taking tests as untimed power tests, but the rest of the class is not. In this way, students with disabilities are treated differently.
Students who use extended-time accommodations are treated differently in another way as well. To implement extended-time accommodations, most universities physically segregate students with disabilities away from their peers, in separate testing locations (Liasidou, 2014). Disabled students are literally excluded from their classrooms. Constitutional Law professor Colker (2019) argues that such segregation violates the spirit, if not the letter, of U.S. disability laws. A more inclusive campus would offer all students the opportunity to experience examinations as untimed power tests in the same room.
In the testing accommodations world, accommodations are frequently distinguished from modifications (Thurlow, Lazarus, Thompson, & Morse, 2005), which is important when evaluating extended time as an accommodation. An accommodation, in contrast to a modification, must not “alter the construct measured” (Sireci, Scarpati, & Li, 2005, p. 460); must only “alter … an unimportant aspect of the test administration” (Lovett, 2014, p. 81); must only be “incidental to the construct being measured” (Brooks, 2003, p. 4); must “only change how a construct is being measured, with the goal of more valid assessment of students with disabilities” (Stretch & Osborne, 2005, p. 2); and must only “increase access to the test content while allowing for the score to be interpreted in the same manner as that for a test taken without an accommodation” (American Educational Research Association, American Psychological Association, and National Council on Measurement in Education, as cited in Cawthon & Leppo, 2013, p. 364).
Therein lies the rub. For extended time to fit the definition of an accommodation rather than a modification, it must not alter the construct measured. Put another way: Standard time limits must, by definition, be only peripheral to the construct measured, immaterial to the test, and incidental to the assessment, in the same way that presenting a test in print rather than Braille would be. The vast timed-testing literature we reviewed above tells us that is certainly not the case. Therefore, we recommend that time limits be removed for all students for all power tests used in educational settings.
Time-Limited Tests Are Less Equitable
Extended time on time-limited tests is a frequent disability accommodation (Casey, 1987; Gregg & Nelson, 2012; Lombardi, Murray, & Gerdes, 2012; Newman, Wagner, Cameto, & Knokey, 2009). In a recent audit, the United States Government Accountability Office (2012) reported that the vast majority of test-related accommodations were for extended time. Extended time on time-limited tests is also considered a “reasonable accommodation,” at least to the degree that most faculty are willing to provide the accommodation.
For example, of the 255 surveyed members of the Association on Handicapped Student Service Programs in Post-Secondary Education, 98% agreed that extended time was a reasonable accommodation; in contrast, fewer than 1% agreed that allowing proofreaders to replace less with more sophisticated vocabulary in students’ draft papers was a reasonable accommodation (Bumba & Goodin, 1986). Of the 420 faculty surveyed at a large public university, 74% reported being very willing to provide extended time; in contrast, only 27% reported being very willing to allow disabled students to complete alternate assessments (e.g., writing a paper, making a presentation, or doing a project in place of a time-limited test; Vogel, Leyser, Wyland, & Brulle, 1999).
Of the 107 faculty surveyed at a small college, 85% reported they would allow disabled students extended time on their exams, but only 33% reported they would allow disabled students to earn extra credit that was not offered to nondisabled students (Nelson, Dodd, & Smith, 1990). Of the 171 faculty surveyed at a research university, 97% reported giving disabled students extended time on their exams, but only 12% reported giving disabled students partial credit for otherwise wrong exam answers (Smith, 2007; see also Houck, Asselin, Troutman, & Arrington, 1992; Matthews, Anderson, & Skolnick, 1987; Norton, 1997; Skinner, 2007; Sweener, Kundert, May, & Quinn, 2002, for similar data from community colleges, private liberal arts colleges, and other institutions).
Thus, extended time on time-limited tests is requested often and considered reasonable, at least in terms of test administrators’ willingness. But is it equitable? The premise of providing extended time is to enable students with disabilities to demonstrate a potential that would otherwise be thwarted by standard time limits. Therefore, the Maximum-Potential Thesis postulates that extended time should aid only students with disabilities (Zuriff, 2000). However, scores of studies, codified by meta-analysis (e.g., Duncan & Purcell, 2019; Gregg & Nelson, 2012; Zuriff, 2000) and systematic review (e.g., Cawthon & Leppo, 2013; Lovett, 2010; Pitoniak & Royer, 2001; Sireci et al., 2005), show that not to be the case. As we have known for a century: Many students, including those without disabilities, are “relatively inefficient in such timed … tests … [but] are able to do relatively efficient and accurate work when allowed to work more slowly” (Mudge, 1921, p. 161).
An alternative to the Maximum-Potential Thesis is the Differential-Boost Hypothesis, which postulates that students without disabilities might be aided by extending or removing time limits, but students with disabilities should be aided more (Fuchs & Fuchs, 2001). Again, numerous studies show that not to be the case. Rather, numerous studies show that removing time limits boosts the performance of numerous students, including students who are learning English (Mullane & McK-elvie, 2001; Pennock-Roman & Rivera, 2011; Solórzano, 2008; Wilson, 1989; Young & King, 2008), students from underrepresented backgrounds (Ardila, 2005; Knapp, 1960), and students who are older than average (Cattell, 1943; Foos & Boone, 2008). Removing time limits also attenuates stereotypic gender differences (Camarata & Woodcock, 2006; De Paola & Gioia, 2016).
Time-limited tests are also less equitable to students with disabilities who forego obtaining documentation and are, therefore, ineligible to receive extended-time accommodations. A wealth of data point to the fact that many students with less visible disabilities, such as anxiety, depression, chronic health conditions (e.g., Crohn’s disease, lupus), dyslexia, and other learning disabilities forego obtaining instructional accommodations because of stigma (Barnard-Brak, Lectenberger, & Lan, 2010; Cole & Cawthon, 2015; Collins & Mowbray, 2005; Denhart, 2008; Ebo, 2016; Grimes, Southgate, Scevak, & Buchanan, 2019; Jaine & Meeks, 2017; Kendall, 2016; Kranke, Jackson, Taylor, Anderson-Fye, & Floersch, 2013; Lightner, Kipps-Vaughan, Schulte, & Trice, 2012; Lyman et al., 2016; Marshak, Van Wieren, Ferrell, Swiss, & Dugan, 2010; Mortimore & Crozier, 2006; Salzer, Wick, & Rogers, 2008; Stanley & Manthorpe, 2001; Stein, 2013).
The stigma of disability is particularly acute for military veterans (Kranke, Weiss, & Constanine Brown, 2017), immigrants (Nadeem et al., 2007), first-generation students (Stebleton, Soria, & Huesman, 2014), and members of cultural minority groups (e.g., Blacks: Ward, Wiltshire, Detry, & Brown, 2013; Asian Americans: Saetermoe, Scat-tone, & Kim, 2001; Latino men: McDonald, Keys, & Balcazar, 2007).
Some students with disabilities also forego obtaining instructional accommodations because of the exorbitant expense of obtaining the required documentation (Grimes et al., 2019; Lightner et al., 2012). As the United States Government Accountability Office (2012) reports, the expense of securing the professional evaluations necessary for disability documentation can amount to several thousand dollars. These costs are almost always borne by the student (or their family) rather than the university (Denhart, 2008; Shipp, 2008), leading to substantial economic barriers for low- and middle-income students and creating disturbing inequities in who receives instructional accommodations (Colker, 2019; Gormley, Hughes, Block, & Lend-mann, 2005; Ragosta, 1987; Ragosta & Wendler, 1992; Wolanin & Steele, 2004).
As the 2019 college entrance scandal vividly illustrates, students from affluent families are considerably more likely to obtain extended-time accommodations. For example, students from affluent public high schools (where only a slim minority of students qualify for free or reduced-cost school lunches) are nearly three times more likely to receive extended-time on college entrance exams than students from low-income public schools (where the majority qualify for free or reduced cost lunches; Belkin, Levitz, & Korn, 2019). Across the nation, students whose families can afford more expensive private college tuition are more likely to have extended-time accommodations (Vickers, 2010). Even at lower cost state universities, students whose families can afford more expensive out-of-state tuition are more likely to have extended-time accommodations (McGregor et al., 2016; see also California State Auditor, 2000; Griggins, 2005; Lerner, 2004).
These inequities could be attenuated by applying Lovett’s (2010, p. 616) recommendation: “if at least some nondisabled examinees would also benefit from the accommodation … it would be inappropriate to withhold the accommodation from them.” Put another way, perhaps it should be instructors who are required to justify, with documentation, why they administer time-limited tests rather than students who are required to justify, with documentation, why they need extended time (Colker, 2019), which leads to our suggestion for further research.
Further Research
Why, nearly a century later, do we continue to “hear school men very authoritatively saying that the fast students make the best grades and the slow ones the poorest?” and why do instructors continue to erroneously assume “that if a student knows the subject in which he is being tested it should follow that he requires but a short time to make his answer?” (Longstaff & Porter, 1928, p. 638). As instructors, we are not assessed by how speedily we can demonstrate our mastery—for instance, how quickly we can teach a semester’s worth of psychology. Why then do we continue to assess our students by how speedily they can demonstrate their mastery? Why do we continue to administer time-limited exams, which are less valid, less reliable, less inclusive, and less equitable? These are questions that summon further research.
We have heard some instructors claim that time-limited tests are needed to ensure fluency in a subject. These instructors often point to math and second language learning as arenas for which time-limited tests are appropriate for building or assessing fluency. However, scholars of math and language learning refute these assumptions.
For example, in a commentary in Education Week, Stanford professor Boaler (2012) chastises teachers and administrators for falsely assuming that mathematical fluency can be developed through time-limited testing. Rather, mathematical fluency and the ability to work facilely with numbers arises from unhurriedly acquiring conceptual knowledge not from anxiously enduring rapid-fire drilling (Boaler, 2014, 2018–2019). Many math educators question whether time-limited testing can even assess, much less develop, fluency (Clarke, Nelson, & Shanley, 2016); some directly deem it ineffectual (Kling & Bay-Williams, 2015), the source of the country’s lackluster mathematical achievement (Flynn, 2017; Pink, 2018), a ritualistic holdover from teachers’ own flawed education (McCloskey, 2014), driven by an erroneous conception of what mathematical skill acquisition is all about (Seeley, 2016). Consequently, the National Council of Teachers of Mathematics (2014) caution against the use of time-limited tests for acquiring or demonstrating mathematical fluency.
Turning to second language instruction, the debate over time-limited versus untimed tests in second language instruction dates as far back as any of the topics we have covered in this article (Carroll & Sapon, 1959, as cited in Carroll, 1990). Even in those earliest days, second language instructors questioned rote drilling as a means for developing fluency (Crawford & Leitzell, 1930, as cited in Nadal, 1950), and researchers demonstrated that fluency was independent of speed (Carroll, 1958). Contemporary educators continue those cautions: When acquiring a second language, speed does not equal proficiency (Laufer & Nation, 2001), and speeded responding should not serve as a proxy for fluency (Segalowitz, 2003) or strength of association (Hemàndez-Chávez, Burt, & Dulay, 1978).
In summary, future research should investigate why instructors continue to cleave to the false beliefs that fluency is evoked and revealed by time-limited testing.
Practice-Based Recommendations
Because time-limited tests are less valid, less reliable, less inclusive, and less equitable, we offer the following recommendations for avoiding time-limited tests in higher educational assessment:
Remove all time limits from all higher educational tests intended to assess power. In addition to improving the tests’ validity, reliability, inclusivity, and equitability, removing time limits from power tests allows students to attenuate their anxiety (Faust, Ashcraft, & Fleck, 1996; Powers, 1986), increase their creativity (Acar & Runco, 2019; Cropley, 1972), read instructions more closely (Myers, 1960), check their work more carefully (Benjamin, Cavell, & Shallenberger, 1984), and learn more thoroughly from prior testing (Chuderski, 2016).
-
If administrative constraints obviate removing time limits (e.g., a testing room cannot be reserved for an unlimited period of time3), administer untimed asynchronous tests (such as take-home exams or untimed online exams). Allow students to work at their own pace and to take advantage of course materials, as well as online information.
If instructors are worried that unlimited testing time in unproctored testing situations (take-home or online exams) will enable students to “look up the answers,” it might be the questions, rather than the students, that need examination. As advised by award-winning educator Geurin (2019), if we are worried about students accessing the Internet to complete exams, perhaps our exams are not asking the right questions: “Google doesn’t have answers. It has information. Students find answers through good thinking.”
As astrophysicist Mack (2018) explains, “A surprisingly large part of having expertise in a topic is not so much knowing everything about it but learning the language and sources well enough to be extremely efficient in Google searches” (see also Boyle, 2018; Coleman, 2019; Edmiston, 2019; Rohrer, 2019). Similarly, as Gernsbacher (2014, p. 3) cautions, “if the answer to a test question is just a click away,” it might not be a very good test question.
Assess mastery through non-test mechanisms, such as projects, reflections, and other performative and demonstrative means of assessment. Gernsbacher’s undergraduate “Research Methods” (https://online225.psych.wisc.edu) and “Psychological Effects of the Internet” (https://Internet.psych.wisc.edu) courses illustrate that a large-enrollment foundational course and an upper-division undergraduate survey course can be conducted with non-test mechanisms of assessment that better parallel the non-test mechanisms of assessment universally present in post-college life.
What is the significance of this article for the general public?
This article cautions instructors against administering tests that have time limits because time-limited tests are less valid, less reliable, less inclusive, and less equitable. Instead, instructors are encouraged to administer untimed power tests, which are tests without any time limits or with time limits so generous that all students are assured of completing all items.
Footnotes
All data and materials supporting the conclusions drawn in this article are available in Gernsbacher, Soicher, and Becker-Blease (2019), which is a Technical Report available on Open Science Framework (https://osf.io/em359).
Although some scholars cite Thorndike (1914) as support for the assumption that faster students perform better, Thorndike’s study of college students computing simple addition problems does not provide this support. The accuracy of the fastest students did not differ from the accuracy of the slowest students.
If the testing room is no longer available, consider allowing students to return to the instructor’s office to continue working on their tests.
Contributor Information
Morton Ann Gernsbacher, Department of Psychology, University of Wisconsin-Madison;.
Raechel N. Soicher, School of Psychological Science, Oregon State University.
Kathryn A. Becker-Blease, School of Psychological Science, Oregon State University.
References
- Abrams EZ, & Goodman JF (1998). Diagnosing developmental problems in children: Parents and professionals negotiate bad news. Journal of Pediatric Psychology, 23, 87–98. 10.1093/jpepsy/23.2.87 [DOI] [PubMed] [Google Scholar]
- Acar S, & Runco MA (2019). Divergent thinking: New methods, recent research, and extended theory. Psychology of Aesthetics, Creativity, and the Arts, 13, 153–158. 10.1037/aca0000231 [DOI] [Google Scholar]
- Ackerman PL, & Ellingsen VJ (2016). Speed and accuracy indicators of test performance under different instructional conditions: Intelligence correlates. Intelligence, 56, 1–9. 10.1016/j.intell.2016.02.004 [DOI] [Google Scholar]
- Anastasi A, & Drake JD (1954). An empirical comparison of certain techniques for estimating the reliability of speeded tests. Educational and Psychological Measurement, 14, 529–540. 10.1177/001316445401400308 [DOI] [Google Scholar]
- Angoff WH (1953). Test reliability and effective test length. Psychometrika, 18, 1–14. 10.1007/BF02289023 [DOI] [Google Scholar]
- Ardila AI (2005). Cultural values underlying psychometric cognitive testing. Neuropsychology Review, 15, 185–195. 10.1007/s11065-005-9180-y [DOI] [PubMed] [Google Scholar]
- Armitage C (1999). Extended time limits on university examinations (Unpublished masters’ thesis). University of Calgary, Calgary, Alberta. [Google Scholar]
- Barnard-Brak L, Lectenberger D, & Lan WY (2010). Accommodation strategies of college students with disabilities. Qualitative Report, 15, 411–429. [Google Scholar]
- Beaulieu RP, & Frost B (1994). Another look at the time-score relationship. Perceptual and Motor Skills, 78, 40–42. 10.2466/pms.1994.78.1.40 [DOI] [Google Scholar]
- Becker MA, & Suls J (1982). Test performance as a function of the hard-driving and speed components of the type A coronary-prone behavior pattern. Journal of Psychosomatic Research, 26, 435–440. 10.1016/0022-3999(82)90018-6 [DOI] [PubMed] [Google Scholar]
- Belkin D, Levitz J, & Korn M (2019, May 21). Many more students, especially the affluent, get extra time to take the SAT; responding to parent pleas, high schools grant special test-taking accommodations to growing numbers, a Wall Street Journal analysis shows. Wall Street Journal. https://www.wsj.com/articles/many-more-students-especially-the-affluent-get-extra-time-to-take-the-sat-11558450347 [Google Scholar]
- Benjamin LT Jr., Cavell TA, & Shallenberger WR III. (1984). Staying with initial answers on objective tests: Is it a myth? Teaching of Psychology, 11, 133–141. 10.1177/009862838401100303 [DOI] [Google Scholar]
- Bennett GK, Seashore HG, & Wesman AG (1956). The differential aptitude tests: An overview. Personnel and Guidance Journal, 35, 81–91. 10.1002/j.2164-4918.1956.tb01710.x [DOI] [Google Scholar]
- Berliner DC (1994). Expertise: The wonders of exemplary performance In Mangieri JN & Block CC (Eds.), Creating powerful thinking in teachers and students (pp. 141–186). Ft. Worth, TX: Holt, Rinehart and Winston. [Google Scholar]
- Boaler J (2012, July 3). Timed tests and the development of math anxiety: Research links ‘torturous’ timed testing to underachievement in math. Education Week. https://www.edweek.org/ew/articles/2012/07/03/36boaler.h31.html [Google Scholar]
- Boaler J (2014). Research suggests that timed tests cause math anxiety. Teaching Children Mathematics, 20, 469–474. 10.5951/teacchilmath.20.8.0469 [DOI] [Google Scholar]
- Boaler J (2018. –2019). Developing mathematical mindsets: The need to interact with numbers flexibly and conceptually. American Educator, 40, 28–33. [Google Scholar]
- Boyle S [@stephanieboyle]. (2018, October 23). I’ve been programming for about 5 years and today googled ‘how to transpose a dataset’, because I hadn’t done it for 2 weeks and couldn’t remember … Coding isn’t about memorising; if it was, I definitely wouldn’t have a job [Tweet]. Twitter. https://twitter.com/_stephanieboyle/status/1054811132324732934 [Google Scholar]
- Bridges KR (1985). Test-completion speed: Its relationship to performance on three course-based objective examinations. Educational and Psychological Measure ment, 45, 29–35. 10.1177/0013164485451003 [DOI] [Google Scholar]
- Brooks TE (2003). Timed versus untimed testing conditions and student performance Assessment Report. San Antonio, TX: Harcourt Educational Measurement. [Google Scholar]
- Brothen T (2012). Time limits on tests: Updating the 1-minute rule. Teaching of Psychology, 39, 288–292. 10.1177/0098628312456630 [DOI] [Google Scholar]
- Bumba R, & Goodin S (1986). Appropriateness of academic adjustments for students with learning disabilities: Perceptions of service providers. AHSSPPE Bulletin, 4, 103–108. [Google Scholar]
- Burack B (1967). Relationship between course examination scores and time taken to finish the examination, revisited. Psychological Reports, 20, 164 10.2466/pr0.1967.20.1.164 [DOI] [PubMed] [Google Scholar]
- Cahalan-Laitusis C, King TC, Cline F, & Bridgeman B (2006). Observational timing study on the SAT Reasoning Test™ for test-takers with learning disabilities and/or AD/HD (College Board Research Report No. 2006–4). New York: NY: College Board Publications. [Google Scholar]
- California State Auditor. (2000). Standardized tests: Although some students may receive extra time on standardized tests that is not deserved, others may not be getting the assistance they need (pp. 2000–2108). Sacramento, CA: Bureau of State Audits. [Google Scholar]
- Camarata S, & Woodcock R (2006). Sex differences in processing speed: Developmental effects in males and females. Intelligence, 34, 231–252. 10.1016/j.intell.2005.12.001 [DOI] [Google Scholar]
- Carroll JB (1958). A factor analysis of two foreign language aptitude batteries. Journal of General Psychology, 59, 3–19. 10.1080/00221309.1958.9710168 [DOI] [PubMed] [Google Scholar]
- Carroll JB (1990). Cognitive abilities in foreign language aptitude: Then and now In Parry TS & Stans-field CW (Eds.), Language aptitude reconsidered (pp. 11–29). Englewood Cliffs, NJ: Prentice Hall. [Google Scholar]
- Carroll JB, & Sapon SM (1959). Modern language aptitude test (MLAT). New York, NY: The Psychological Corporation. [Google Scholar]
- Carter WR (1932). A study of certain mathematical abilities in high school physics. Mathematics Teacher, 25, 388–419. [Google Scholar]
- Casey E (1987). Accommodating testing to disabled students In Bray D & Belcher MJ (Eds.), New directions for community college, No. 59 (pp. 65–73). San Francisco, CA: Jossey-Bass. [Google Scholar]
- Cattell RB (1943). The measurement of adult intelligence. Psychological Bulletin, 40, 153–193. 10.1037/h0059973 [DOI] [Google Scholar]
- Cawthon S, & Leppo R (2013). Assessment accommodations on tests of academic achievement for students who are deaf or hard of hearing: A qualitative meta-analysis of the research literature. American Annals of the Deaf, 158, 363–376. 10.1353/aad.2013.0023 [DOI] [PubMed] [Google Scholar]
- Centra JA (1986). Handicapped student performance on the scholastic aptitude test. Journal of Learning Disabilities, 19, 324–327. 10.1177/002221948601900602 [DOI] [PubMed] [Google Scholar]
- Chuderski A (2016). Time pressure prevents relational learning. Learning and Individual Differences, 49, 361–365. 10.1016/j.lindif.2016.07.006 [DOI] [Google Scholar]
- Clarke B, Nelson N, & Shanley L (2016). Mathematics fluency—More than the weekly timed test In Cummings KD & Petscher Y (Eds.), The fluency construct (pp. 67–89). New York, NY: Springer; 10.1007/978-1-4939-2803-3_3 [DOI] [Google Scholar]
- Cole EV, & Cawthon SW (2015). Self-disclosure decisions of university students with learning disabilities. Journal of Postsecondary Education and Disability, 28, 163–179. [Google Scholar]
- Coleman G [@GraemeColeman]. (2019, February 20). Hello, my name is Graeme, I have a PhD in computing, and I am a senior accessibility consultant, but when I want to type “é” on a Windows laptop I go to Beyoncé’s Wikipedia p. and copy/paste the letter from there [Tweet]. Twitter. https://twitter.com/graemecoleman/status/1098253794691944448 [Google Scholar]
- Colker R (2019). Test validity: Faster is not necessarily better. Seton Hall Law Review, 49, 679–736. [Google Scholar]
- Collins ME, & Mowbray CT (2005). Higher education and psychiatric disabilities: National survey of campus disability services. American Journal of Orthopsychiatry, 75, 304–315. 10.1037/0002-9432.75.2.304 [DOI] [PubMed] [Google Scholar]
- Cooper PL (1984). The assessment of writing ability: A review of research (GRE Board Research Report No. 82–15R and ETS Research Report 84–12). Princeton, NJ: Educational Testing Service. [Google Scholar]
- Crawford CC, & Leitzell E (1930). Learning a new language. Los Angeles, CA: C. C. Crawford. [Google Scholar]
- Crocker L, & Algina J (1986). Introduction to classical and modern test theory. New York, NY: Holt, Rinehart, and Winston. [Google Scholar]
- Cronbach LJ (1949). Essentials of psychological testing. New York, NY: Harper and Brothers. [Google Scholar]
- Cronbach LJ, & Warrington WG (1951). Time-limit tests: Estimating their reliability and degree of speeding. Psychometrika, 16, 167–188. 10.1007/BF02289113 [DOI] [PubMed] [Google Scholar]
- Cropley AJ (1972). Originality scores under timed and untimed conditions. Australian Journal of Psychology, 24, 31–36. 10.1080/00049537208255782 [DOI] [Google Scholar]
- Danthiir V, Wilhelm O, & Schacht A (2005). Decision speed in intelligence tasks: Correctly an ability? Psychology Science, 47, 200–229. [Google Scholar]
- Denhart H (2008). Deconstructing barriers: Perceptions of students labeled with learning disabilities in higher education. Journal of Learning Disabilities, 41, 483–497. 10.1177/0022219408321151 [DOI] [PubMed] [Google Scholar]
- De Paola M, & Gioia F (2016). Who performs better under time pressure? Results from a field experiment. Journal of Economic Psychology, 53, 37–53. 10.1016/j.joep.2015.12.002 [DOI] [Google Scholar]
- Dodonova YA, & Dodonov YS (2013). Faster on easy items, more accurate on difficult ones: Cognitive ability and performance on a task of varying difficulty. Intelligence, 41, 1–10. 10.1016/j.intell.2012.10.003 [DOI] [Google Scholar]
- Donlon TF (1975, March-April). Estimating rate-of-work parameters in the assessment of test speededness. Paper presented at the Annual Meeting of the National Council on Measurement in Education Washington, DC. [Google Scholar]
- Donlon TF (1984). The Scholastic Aptitude Test In Donlon TF (Ed.), The College Board technical handbook for the Scholastic Aptitude Test and Achievement Tests (pp. 37–67). New York, NY: College Entrance Examination Board. [Google Scholar]
- DuBois PH (1932). A speed factor in mental tests In Woodworth RS (Ed.), Archives of psychology (Vol. 141, pp. 1–38). New York, NY: Publisher Unknown. [Google Scholar]
- Duncan H, & Purcell C (2019). Consensus or contradiction? A review of the current research into the impact of granting extra time in exams to students with specific learning difficulties (SpLD). Journal of Further and Higher Education. Advance online publication; 10.1080/0309877X.2019.1578341 [DOI] [Google Scholar]
- Ebel RL, & Frisbie DA (1991). Essentials of educational measurement (5th ed.). New Delhi, India: Prentice Hall of India. [Google Scholar]
- Ebo BA (2016). Understanding the experiences of college students with learning disabilities. (Unpublished doctoral dissertation). Northeastern University, Boston, MA. [Google Scholar]
- Edmiston EK [@EKaleEdmiston]. (2019, January 4). A friend/colleague who is an excellent programmer offhandedly told me the other day that coding is 90% googling error messages & 10% writing code. Until this point, I thought that all the time I spent [Tweet]. Twitter. https://twitter.com/ekaleedmiston/status/1081221822186696706 [Google Scholar]
- Elliott SN, McKevitt BC, & Kettler RJ (2002). Testing accommodations research and decision making: The case of “good” scores being highly valued but difficult to achieve for all students. Measurement and Evaluation in Counseling and Development, 35, 153–166. 10.1080/07481756.2002.12069060 [DOI] [Google Scholar]
- Faust MW, Ashcraft MH, & Fleck DE (1996). Mathematics anxiety effects in simple and complex addition. Mathematical Cognition, 2, 25–62. 10.1080/135467996387534 [DOI] [Google Scholar]
- Flynn M (2017). Beyond answers: Exploring mathematical practices with young children. Portsmouth, NH: Sten-house Publishers. [Google Scholar]
- Foos PW (1989). Completion time and performance on multiple-choice and essay tests. Bulletin of the Psycho-nomic Society, 27, 179–180. 10.3758/BF03329933 [DOI] [Google Scholar]
- Foos PW, & Boone D (2008). Adult age differences in divergent thinking: It’s just a matter of time. Educational Gerontology, 34, 587–594. 10.1080/03601270801949393 [DOI] [Google Scholar]
- Freeman FN (1923). Note on the relation between speed and accuracy or quality of work. The Journal of Educational Research, 7, 87–90. 10.1080/00220671.1923.10879332 [DOI] [Google Scholar]
- Freeman FS (1932). The factor of speed. Journal of General Psychology, 6, 462–468. 10.1080/00221309.1932.9711887 [DOI] [Google Scholar]
- Fuchs LS, & Fuchs D (2001). Helping teachers formulate sound test accommodation decisions for students with learning disabilities. Learning Disabilities Research & Practice, 16, 174–181. 10.1111/0938-8982.00018 [DOI] [Google Scholar]
- Gernsbacher MA (2015). Diverse brains. The General Psychologist, 49, 29–37. [PMC free article] [PubMed] [Google Scholar]
- Gernsbacher MA (2014). Why internet-based education? Frontiers in Psychology, 5, 1530 10.3389/fpsyg.2014.01530 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gernsbacher MA, Soicher RN, & Becker-Blease KA (2019). Arguments against time-limited educational testing: A technical report. Open Science Framework. [Google Scholar]
- Geurin D [@DavidGeurin]. (2019, September 9). If students are able to rely on the Internet to get answers, are we asking the right questions? Google doesn’t have answers. It has information. Students find answers through good thinking [Tweet]. Twitter. https://twitter.com/DavidGeurin/status/1171204304159612928. [Google Scholar]
- Ghaffari S (2015). Investigating the relationship between student performance and average time spent on chemistry exams. AURCO Journal, 21, 72–80. http://www.aurco.org/Journals/AURCO_Journal_2015/Ghaffari.pdf [Google Scholar]
- Goode E (2001, May 16). Anne Anastasi, the ‘test guru’ of psychology, is dead at 92. The New York Times. https://www.nytimes.com/2001/05/16/nyregion/anne-anastasithe-test-guru-of-psychology-is-dead-at-92.html [Google Scholar]
- Gormley S, Hughes C, Block L, & Lendmann C (2005). Eligibility assessment requirements at the postsecondary level for students with learning disabilities: A disconnect with secondary schools? Journal of Postsecondary Education and Disability, 18, 63–70. [Google Scholar]
- Gregg N, & Nelson JM (2012). Meta-analysis on the effectiveness of extra time as a test accommodation for transitioning adolescents with learning disabilities: More questions than answers. Journal of Learning Disabilities, 45, 128–138. 10.1177/0022219409355484 [DOI] [PubMed] [Google Scholar]
- Griggins C (2005). Dosing dilemmas: Are you rich and white or poor and black? The American Journal of Bioethics, 5, 55–57. 10.1080/15265160591002782 [DOI] [PubMed] [Google Scholar]
- Grimes S, Southgate E, Scevak J, & Buchanan R (2019). University student perspectives on institutional non-disclosure of disability and learning challenges: Reasons for staying invisible. International Journal of Inclusive Education, 23, 639–655. 10.1080/13603116.2018.1442507 [DOI] [Google Scholar]
- Hadley WM (2007). The necessity of academic accommodations for first-year college students with learning disabilities. Journal of College Admission, 195, 9–13. https://files.eric.ed.gov/fulltext/EJ783943.pdf [Google Scholar]
- Hammonds F, & Mariano G (2015). Student test grades in college: A study of possible predictors. International Journal on Teaching and Learning in Higher Education, 27, 114–118. https://files.eric.ed.gov/fulltext/EJ1069824.pdf [Google Scholar]
- Henderson WD (2005). The LSAT, law school exams, and meritocracy: The surprising and under-theorized role of test-taking speed. Texas Law Review, 82, 975–1051. [Google Scholar]
- Herman WE (1997). The relationship between time to completion and achievement on multiple-choice exams. Journal of Research & Development in Education, 30, 113–117. [Google Scholar]
- Hemàndez-Chávez E, Burt M, & Dulay H (1978). Language dominance and proficiency testing: Some general considerations. NABE Journal, 3, 41–54. 10.1080/08855072.1978.10668343 [DOI] [Google Scholar]
- Hishinuma ES (1998). Issues related to WAIS-R testing modifications for individuals with learning disabilities or Attention-Deficit/Hyperactivity Disorder. Learning Disability Quarterly, 21, 228–240. 10.2307/1511084 [DOI] [Google Scholar]
- Holmes A, & Silvestri R (2019). Extra time or unused time? What data from a college testing center tells us about 50% extra time as an accommodation for students with learning disabilities. Psychological Injury and Law, 12, 7–16. 10.1007/s12207-019-09339-9 [DOI] [Google Scholar]
- Houck CK, Asselin SB, Troutman GC, & Arrington JM (1992). Students with learning disabilities in the university environment: A study of faculty and student perceptions. Journal of Learning Disabilities, 25, 678–684. 10.1177/002221949202501008 [DOI] [PubMed] [Google Scholar]
- Jaine NR, & Meeks LM (2017). Privacy, disability, and health science students. Disability Compliance for Higher Education, 22, 7 10.1002/dhe.30270 [DOI] [Google Scholar]
- Johnston JJ (1987). Exam-taking speed and grades In Millard RJ & Ware ME (Eds.), Handbook on student development: Advising, career development, and field placement (p. 39). Hillsdale, NJ: Erlbaum. [Google Scholar]
- Kendall L (2016). Higher education and disability: Exploring student experiences. Cogent Education, 3, 1256142 10.1080/2331186X.2016.1256142 [DOI] [Google Scholar]
- Kennedy R (1994, November). A study of the relationship between scores and time on tests. Paper presented at the Annual Meeting of the Mid-South Educational Research Association, Nashville, TN: https://files.eric.ed.gov/full-text/ED392833.pdf [Google Scholar]
- Kerstiens G (1990). A slow look at speeded reading comprehension tests. Review of Research in Developmental Education, 7, 3. [Google Scholar]
- Kling G, & Bay-Williams JM (2015). Three steps to mastering multiplication facts. Teaching Children Mathematics, 21, 548–559. 10.5951/teacchilmath.21.9.0548 [DOI] [Google Scholar]
- Knapp RR (1960). The effects of time limits on the intelligence test performance of Mexican and American subjects. Journal of Educational Psychology, 51, 14–20. 10.1037/h0038366 [DOI] [Google Scholar]
- Kranke D, Jackson SE, Taylor DA, Anderson-Fye E, & Floersch J (2013). College student disclosure of non-apparent disabilities to receive classroom accommodations. Journal of Postsecondary Education and Disability, 26, 35–51. [Google Scholar]
- Kranke DA, Weiss EL, & Constanine Brown J (2017). Student veterans with invisible disabilities: Accommodation-seeking in higher education. Journal of Veterans Studies, 2, 45–57. 10.21061/jvs.15 [DOI] [Google Scholar]
- Kyllonen PC, & Zu J (2016). Use of response time for measuring cognitive ability. Journal of Intelligence, 4, 14 10.3390/jintelligence4040014 [DOI] [Google Scholar]
- Laitusis CC, Morgan DL, Bridgeman B, Zanna J, & Stone E (2007). Examination of fatigue effects from extended-time accommodations on the SAT Reasoning Test (College Board Research Report No. 2007–1 and ETS RR-07–31). New York: NY: College Board Publications. [Google Scholar]
- Laufer B, & Nation P (2001). Passive vocabulary size and speed of meaning recognition: Are they related? EUROSLA Yearbook, 1, 7–28. 10.1075/eurosla.1.05lau [DOI] [Google Scholar]
- Lee S (2012). Education briefing. Journal of Property Investment & Finance, 30, 6 10.1108/jpif.2012.11230faa.003 [DOI] [Google Scholar]
- Lemmon VW (1927). The relation of reaction time to measures of intelligence, memory, and learning In Woodworth RS (Ed.), Archives of psychology (Vol. 94, pp. 1–38). New York, NY: Publisher Unknown. [Google Scholar]
- Lerner CS (2004). “Accommodations” for the learning disabled: A level playing field or affirmative action for elites? Vanderbilt Law Review, 57, 1043–1124. [Google Scholar]
- Lester D (1991). Speed and performance on college course examinations. Perceptual and Motor Skills, 73, 1090 10.2466/pms.1991.73.3f.1090 [DOI] [Google Scholar]
- Liasidou A (2014). Critical disability studies and socially just change in higher education. British Journal of Special Education, 41, 120–135. 10.1111/1467-8578.12063 [DOI] [Google Scholar]
- Lightner KL, Kipps-Vaughan D, Schulte T, & Trice AD (2012). Reasons university students with a learning disability wait to seek disability services. Journal of Postsecondary Education and Disability, 25, 145–159. [Google Scholar]
- Line W, & Kaplan E (1932). The existence, measurement and significance of a speed factor in the abilities of public school children. Journal of Experimental Education, 1, 1–8. 10.1080/00220973.1932.11009882 [DOI] [Google Scholar]
- Lombardi AR, Murray C, & Gerdes H (2012). Academic performance of first-generation college students with disabilities. Journal of College Student Development, 53, 811–826. 10.1353/csd.2012.0082 [DOI] [Google Scholar]
- Longstaff HP, & Porter JP (1928). Speed and accuracy as factors in objective tests in general psychology. Journal of Applied Psychology, 12, 636–642. 10.1037/h0071593 [DOI] [Google Scholar]
- Lord FM (1956). A study of speed factors in tests and academic grades. Psychometrika, 21, 31–50. 10.1007/BF02289085 [DOI] [Google Scholar]
- Lorge I (1936). The influence of the test upon the nature of mental decline as a function of age. Journal of Educational Psychology, 27, 100–110. 10.1037/h0059078 [DOI] [Google Scholar]
- Lovett BJ (2010). Extended time testing accommodations for students with disabilities: Answers to five fundamental questions. Review of Educational Research, 80, 611–638. 10.3102/0034654310364063 [DOI] [Google Scholar]
- Lovett BJ (2014). Testing accommodations under the amended Americans with Disabilities Act: The voice of empirical research. Journal of Disability Policy Studies, 25, 81–90. 10.1177/1044207312469830 [DOI] [Google Scholar]
- Lovett BJ, Lewandowski LJ, & Potts HE (2017). Test-taking speed: Predictors and implications. Journal of Psychoeducational Assessment, 35, 351–360. 10.1177/0734282916639462 [DOI] [Google Scholar]
- Lu Y, & Sireci SG (2007). Validity issues in test speededness. Educational Measurement: Issues and Practice, 26, 29–37. 10.1111/j.1745-3992.2007.00106.x [DOI] [Google Scholar]
- Lyman M, Beecher ME, Griner D, Brooks M, Call J, & Jackson A (2016). What keeps students with disabilities from using accommodations in postsecondary education? A qualitative review. Journal of Postsecondary Education and Disability, 29, 123–140. [Google Scholar]
- Mack K [@AstroKatie]. (2018, December 8). A surprisingly large part of having expertise in a topic is not so much knowing everything about it but learning the language and sources well enough to be extremely efficient in google searches [Tweet]. Twitter. https://twitter.com/AstroKatie/status/1071442842873159681 [Google Scholar]
- Mangan GL (1959). A factorial study of speed, power and related temperament variables. The British Journal of Educational Psychology, 29, 144–154. 10.1111/j.2044-8279.1959.tb01488.x [DOI] [Google Scholar]
- Marshak L, Van Wieren T, Ferrell D, Swiss L, & Dugan C (2010). Exploring barriers to college student use of disability services and accommodations. Journal of Postsecondary Education and Disability, 22, 151–165. [Google Scholar]
- Matthews PR, Anderson DW, & Skolnick BD (1987). Faculty attitude toward accommodations for college students with learning disabilities. Learning Disabilities Focus, 3, 46–52. [Google Scholar]
- McCall WA (1916). Correlation of some psychological and educational measurements, with special attention to the measurement of mental ability. New York, NY: Columbia University Teachers College. [Google Scholar]
- McCloskey A (2014). The promise of ritual: A lens for understanding persistent practices in mathematics classrooms. Educational Studies in Mathematics, 86, 19–38. 10.1007/s10649-013-9520-4 [DOI] [Google Scholar]
- McDannell L, & Lyvers Peffer PAL (2013). The relationship between exam completion time and exam score in an introductory animal science course (Unpublished senior honors thesis). The Ohio State University, Columbus, OH: https://kb.osu.edu/handle/1811/54580 [Google Scholar]
- McDonald KE, Keys CB, & Balcazar FE (2007). Disability, race/ethnicity and gender: Themes of cultural oppression, acts of individual resistance. American Journal of Community Psychology, 39, 145–161. 10.1007/s10464-007-9094-3 [DOI] [PubMed] [Google Scholar]
- McGregor KK, Langenfeld N, Van Horne S, Oleson J, Anson M, & Jacobson W (2016). The university experiences of students with learning disabilities. Learning Disabilities Research & Practice, 31, 90–102. 10.1111/ldrp.12102 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mead AD, & Drasgow F (1993). Equivalence of computerized and paper-and-pencil cognitive ability tests: A meta-analysis. Psychological Bulletin, 114, 449–458. 10.1037/0033-2909.114.3.449 [DOI] [Google Scholar]
- Merriam-Webster. (n.d.). Retardation. Merriam Webster Dictionary. https://www.merriam-webster.com/dictionary/retardation [Google Scholar]
- Michael JJ, & Michael WB (1969). The relationship of performance on objective examinations to the order in which students complete them. Educational and Psychological Measurement, 29, 511–513. 10.1177/001316446902900233 [DOI] [Google Scholar]
- Miller TW, & Weiss DJ (1976). Effects of time limits on test-taking behavior (Research Report 76–2). Minneapolis, MN: University of Minnesota. [Google Scholar]
- Mollenkopf WG (1950). Slow—But how sure? College Board Review, 11, 147–151. [Google Scholar]
- Mollenkopf WG (1960). Time limits and the behavior of test takers. Educational and Psychological Measurement, 20, 223–230. 10.1177/001316446002000203 [DOI] [Google Scholar]
- Morrison EJ (1960). On test variance and the dimensions of the measurement situation. Educational and Psychological Measurement, 20, 231–250. 10.1177/001316446002000204 [DOI] [Google Scholar]
- Mortimore T, & Crozier WR (2006). Dyslexia and difficulties with study skills in higher education. Studies in Higher Education, 31, 235–251. 10.1080/03075070600572173 [DOI] [Google Scholar]
- Mudge EL (1921). Time and accuracy as related to mental tests. Journal of Educational Psychology, 12, 159–161. 10.1037/h0069817 [DOI] [Google Scholar]
- Mullane J, & McKelvie SJ (2001). Effects of removing the time limit on first and second language intelligence test performance. Practical Assessment, Research & Evaluation, 7, 1–6. https://doaj.org/article/fb413fc32cfd43068694d1c57854e50c [Google Scholar]
- Myers CT (1960). Introduction. Educational and Psychological Measurement, 20, 221–222. 10.1177/001316446002000202 [DOI] [Google Scholar]
- Nadal RK (1950). Summary of research on methods and devices for teaching the vocabulary of English as a foreign language. (Unpublished master’s thesis). Boston University, Boston, MA. [Google Scholar]
- Nadeem E, Lange JM, Edge D, Fongwa M, Belin T, & Miranda J (2007). Does stigma keep poor young immigrant and U. S.-born Black and Latina women from seeking mental health care? Psychiatric Services, 58, 1547–1554. 10.1176/ps.2007.58.12.1547 [DOI] [PubMed] [Google Scholar]
- National Association of Secondary-School Principals. (1948). Part II: What tests? NASSP Bulletin, 32, 10–55. [Google Scholar]
- National Council of Teachers of Mathematics. (2014). Principles to actions: Ensuring mathematical success for all. Reston, VA: NCTM Board of Directors. [Google Scholar]
- Nelson JR, Dodd JM, & Smith DJ (1990). Faculty willingness to accommodate students with learning disabilities: A comparison among academic divisions. Journal of Learning Disabilities, 23, 185–189. 10.1177/002221949002300309 [DOI] [PubMed] [Google Scholar]
- Neubauer AC (1990). Speed of information processing in the hick paradigm and response latencies in a psychometric intelligence test. Personality and Individual Differences, 11, 147–152. 10.1016/0191-8869(90)90007-E [DOI] [Google Scholar]
- Nevo B, & Spector A (1979). Personal tempo in taking tests of the multiple-choice type. The Journal of Educational Research, 73, 75–78. 10.1080/00220671.1979.10885211 [DOI] [Google Scholar]
- Newman L, Wagner M, Cameto R, & Knokey A-M (2009). The post-high School outcomes of youth with disabilities up to 4 years after high school: A report from the national longitudinal transition study-2 (NLTS2) (NCSER 2009–3017). Menlo Park, CA: SRI International. [Google Scholar]
- Norton SM (1997). Examination accommodations for community college students with learning disabilities: How are they viewed by faculty and students? Community College Journal of Research and Practice, 21, 57–69. 10.1080/1066892970210106 [DOI] [Google Scholar]
- Onwuegbuzie AJ, & Daley CE (1996). The relative contributions of examination-taking coping strategies and study coping strategies to test anxiety: A concurrent analysis. Cognitive Therapy and Research, 20, 287–303. 10.1007/BF02229239 [DOI] [Google Scholar]
- Pardy B (2017, August 17). Mental disabilities shouldn’t be accommodated with extra time on exams. National Post. https://nationalpost.com/opinion/bruce-pardy-mental-disabilities-shouldnt-be-accommodated-with-extra-time-on-exams [Google Scholar]
- Parr P, Levi N, & Jacka P (1996). Unspeeded examinations: An equitable and practical method of assessment. [Technical Report]. https://files.eric.ed.gov/fulltext/ED397108.pdf [Google Scholar]
- Paul CA, & Rosenkoetter JS (1980). The relationship between the time taken to complete an examination and the test score received. Teaching of Psychology, 7, 108–109. [Google Scholar]
- Peak H, & Boring EG (1926). The factor of speed in intelligence. Journal of Experimental Psychology, 9, 71–94. 10.1037/h0071020 [DOI] [Google Scholar]
- Pennock-Roman M, & Rivera C (2011). Mean effects of test accommodations for ELLs and non-ELLs: A meta-analysis of experimental studies. Educational Measurement: Issues and Practice, 30, 10–28. 10.1111/j.1745-3992.2011.00207.x [DOI] [Google Scholar]
- Peterson NG (1993). Review of issues associated with speededness of GATB tests. Washington, DC: American Institutes for Research. [Google Scholar]
- Phillips SE (1994). High-stakes testing accommodations: Validity versus disabled rights. Applied Measurement in Education, 7, 93–120. 10.1207/s15324818ame0702_1 [DOI] [Google Scholar]
- Pink EL (2018). The algebra of empathy. The Cutting Edge: The Stanford Undergraduate Journal of Education Research, 2, 1 http://ojs.stanford.edu/ojs/index.php/ce/article/view/1159 [Google Scholar]
- Pitoniak MJ, & Royer JM (2001). Testing accommodations for examinees with disabilities: A review of psychometric, legal, and social policy issues. Review of Educational Research, 71, 53–104. 10.3102/00346543071001053 [DOI] [Google Scholar]
- Powers DE (1986). Test anxiety and the GRE General Test (GRE Board Professional Report No. 83-l7P). Princeton, NJ: Educational Testing Service. [Google Scholar]
- Ragosta M (1987). Students with disabilities: Four years of data from special test administrations of the scholastic aptitude test 1980–83 (College Board Report No. 87–2 and ETS RR 87–2). New York, NY: College Board Publications. [Google Scholar]
- Ragosta M, & Wendler C (1992). Eligibility issues and comparable time limits for disabled and non-disabled SAT examinees (College Board Report No. 92–5 ETS RR 92–35). New York, NY: College Board Publications. [Google Scholar]
- Rogers BG (1968). The prediction of time scores on achievement tests from academic variables. (Unpublished doctoral dissertation). Michigan State University, East Lansing, MI. [Google Scholar]
- Rohrer B [@_brohrer_]. (2019, October 8). I was just asked whether I ever have to look things up when I code. I want to go on record saying that, aside from canned white board coding examples, I can’t write two lines of code without referring to Google or stack overflow. I would be lost without them [Tweet]. Twitter. https://twitter.com/_brohrer_/status/1049388373541089280 [Google Scholar]
- Ruch GM, & Koerth W (1923). “Power” vs. “speed” in Army Alpha. Journal of Educational Psychology, 14, 193–208. 10.1037/h0074006 [DOI] [Google Scholar]
- Saetermoe CL, Scattone D, & Kim KH (2001). Ethnicity and the stigma of disabilities. Psychology & Health, 16, 699–713. 10.1080/08870440108405868 [DOI] [Google Scholar]
- Salzer MS, Wick LC, & Rogers JA (2008). Familiarity with and use of accommodations and supports among postsecondary students with mental illnesses. Psychiatric Services, 59, 370–375. 10.1176/ps.2008.59.4.370 [DOI] [PubMed] [Google Scholar]
- Schnusenberg O, & Slater R (2011). Do the good students finish first? Journal of Business and Behavioral Sciences, 23, 15–34. [Google Scholar]
- Seeley CL (2016). Building a math-positive culture: How to support great math teaching in your school. Alexandria, VA: Association for Supervision and Curriculum Development. [Google Scholar]
- Segalowitz N (2003). Automaticity and second languages In Doughty CJ & Long MH (Eds.), The handbook of second language acquisition (pp. 382–408). Malden, MA: Blackwell Publishing; 10.1002/9780470756492.ch13 [DOI] [Google Scholar]
- Shipp AE (2008). Disability documentation criteria for students with learning disabilities in higher education. (Unpublished doctoral dissertation). Auburn University, Auburn, AL. [Google Scholar]
- Sireci SG, Scarpati SE, & Li S (2005). Test accommodations for students with disabilities: An analysis of the interaction hypothesis. Review of Educational Research, 75, 457–490. 10.3102/00346543075004457 [DOI] [Google Scholar]
- Sisk TK (1926). The interrelations of speed in simple and complex responses. Nashville, TN: George Peabody College for Teachers. [Google Scholar]
- Skinner ME (2007). Faculty willingness to provide accommodations and course alternatives to post-secondary students with learning disabilities. International Journal of Special Education, 22, 32–45. [Google Scholar]
- Slater P (1938). Speed of work in intelligence tests. British Journal of Psychology, 29, 55–68. [Google Scholar]
- Smith SC (2007). Attitudes towards accommodations for students with learning disabilities: A study of Rowan University faculty. Theses and Dissertations, 845 https://rdw.rowan.edu/etd/845 [Google Scholar]
- Solórzano RW (2008). High stakes testing: Issues, implications, and remedies for English language learners. Review of Educational Research, 78, 260–329. 10.3102/0034654308317845 [DOI] [Google Scholar]
- Spenceley LM, & Wheeler S (2016). The use of extended time by college students with disabilities. Journal of Postsecondary Education and Disability, 29, 141–150. [Google Scholar]
- Stanley N, & Manthorpe J (2001). Responding to students’ mental health needs: Impermeable systems and diverse users. Journal of Mental Health, 10, 41–52. 10.1080/09638230020023606 [DOI] [Google Scholar]
- Stebleton MJ, Soria KM, & Huesman RL Jr. (2014). First-generation students’ sense of belonging, mental health, and use of counseling services at public research universities. Journal of College Counseling, 17, 6–20. 10.1002/j.2161-1882.2014.00044.x [DOI] [Google Scholar]
- Stein KF (2013). DSS and accommodations in higher education: Perceptions of students with psychological disabilities. Journal of Postsecondary Education and Disability, 26, 145–161. [Google Scholar]
- Stretch LAS, & Osborne JW (2005). Extended time test accommodation: Directions for future research and practice. Practical Assessment, Research & Evaluation, 10, 8 10.7275/cs6a-4s02 [DOI] [Google Scholar]
- Sweener K, Kundert D, May D, & Quinn K (2002). Comfort with accommodations at the community college level. Journal of Developmental Education, 25, 12–18. [Google Scholar]
- Terranova C (1972). Relationship between test scores and test time. Journal of Experimental Education, 40, 81–83. [Google Scholar]
- The American Heritage Dictionary of American English. (n.d.). Retard. The American Heritage Dictionary. https://ahdictionary.com/word/search.html?q=retard [Google Scholar]
- Thissen D (1983). Timed testing: An approach using item response theory New horizons in testing: Latent trait test theory and computerized adaptive testing (pp. 179–203). New York, NY: Academic Press. [Google Scholar]
- Thorndike EL (1914). On the relation between speed and accuracy in addition. Journal of Educational Psychology, 5, 537 10.1037/h0075308 [DOI] [Google Scholar]
- Thurlow ML, Lazarus SS, Thompson SJ, & Morse AB (2005). State policies on assessment participation and accommodations for students with disabilities. The Journal of Special Education, 38, 232–240. 10.1177/00224669050380040401 [DOI] [Google Scholar]
- Thurlow ML, Seyfarth AL, Scott DL, & Ysseldyke JE (1997). State assessment policies on participation and accommodations for students with disabilities: 1997 update (Synthesis Report 29). Minneapolis, MN: National Center on Educational Outcomes. [Google Scholar]
- Tinker MA (1944). Speed, power, and level in the revised Minnesota Paper Form Board Test. The Pedagogical Seminary and Journal of Genetic Psychology, 64, 93–97. 10.2307/1416865 [DOI] [Google Scholar]
- Traub RE, & Hambleton RK (1972). The effect of scoring instructions and degree of speededness on the validity and reliability of multiple-choice tests. Educational and Psychological Measurement, 32, 737–758. 10.1177/001316447203200313 [DOI] [Google Scholar]
- Tryon CMG, & Jones HE (1933). The relationship between “speed” and “altitude.” Journal of Experimental Psychology, 16, 98–114. 10.1037/h0072763 [DOI] [Google Scholar]
- United States Department of Justice. (2019, March 11). Affidavit in support of criminal complaint. U.S. Attorney’s Office District of Massachusetts; https://www.justice.gov/file/1142876/download [Google Scholar]
- United States Department of Justice. (2014). ADA requirements: Testing accommodations. Civil Rights Division: Disability Rights Section; www.ada.gov/regs2014/testing_accommodations.html [Google Scholar]
- United States Government Accountability Office. (2012). Higher education and disability: Improved federal enforcement needed to better protect students’ rights to testing accommodations (GAO-12–40). Washington, DC: U.S. Government Accountability Office. [Google Scholar]
- Vernon PA, & Kantor L (1986). Reaction time correlations with intelligence test scores obtained under either timed or untimed conditions. Intelligence, 10, 315–330. 10.1016/0160-2896(86)90002-4 [DOI] [Google Scholar]
- Vickers MZ (2010, March). Accommodating college students with learning disabilities: ADD, ADHD, and dyslexia. Raleigh, NC: Pope Center Series on Higher Education. [Google Scholar]
- Vogel SA, Leyser Y, Wyland S, & Brulle A (1999). Students with learning disabilities in higher education: Faculty attitude and practices. Learning Disabilities Research & Practice, 14, 173–186. 10.1207/sldrp1403_5 [DOI] [Google Scholar]
- Ward EC, Wiltshire JC, Detry MA, & Brown RL (2013). African American men and women’s attitude toward mental illness, perceptions of stigma, and preferred coping behaviors. Nursing Research, 62, 185–194. 10.1097/NNR.0b013e31827bf533 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wesman AG (1949). Effect of speed on item-test correlation coefficients. Educational and Psychological Measurement, 9, 51–57. 10.1177/001316444900900106 [DOI] [Google Scholar]
- Whipple GM (1914). Manual of mental and physical tests. Baltimore, MD: Warwick & York, Inc. [Google Scholar]
- White PO (1973). Individual differences in speed, accuracy and persistence: A mathematical model for problem solving In Eysenck HJ (Ed.), The measurement of intelligence (pp. 246–260). Dordrecht, Germany: Springer; 10.1007/978-94-011-6129-9_16 [DOI] [Google Scholar]
- Wierzbicki M (1994). Relation between order of completion and performance on timed examinations. Psychological Reports, 74, 411–414. 10.2466/pr0.1994.74.2.411 [DOI] [Google Scholar]
- Wilhelm O, & Schulze R (2002). The relation of speeded and unspeeded reasoning with mental speed. Intelligence, 30, 537–554. 10.1016/S0160-2896(02)00086-7 [DOI] [Google Scholar]
- Wilson KM (1989). Population differences in speed versus level of GRE Reading Comprehension: An exploratory study (GRE Board Professional Report No. 84–09aP and ETS Research Report 89–36). Princeton, NJ: Educational Testing Service. [Google Scholar]
- Wolanin TR, & Steele PE (2004). Higher education opportunities for students with disabilities: A primer for policymakers. Washington, DC: The Institute for Higher Education Policy. [Google Scholar]
- Yates AJ (1963). A further study of progressive matrices (1947). British Journal of Educational Psychology, 33, 307–311. 10.1111/j.2044-8279.1963.tb00593.x [DOI] [Google Scholar]
- Yates AJ (1966). The relationship between level and speed on two intelligence tests. British Journal of Educational Psychology, 36, 166–170. 10.1111/j.2044-8279.1966.tb01865.x [DOI] [PubMed] [Google Scholar]
- Young JW, & King TC (2008). Testing accommodations for English Language Learners: A review of state and district policies (College Board Research Report No. 2008–6 and ETS RR 08–48). New York: NY: College Board Publications. [Google Scholar]
- Zuriff GE (2000). Extra examination time for students with learning disabilities: An examination of the maximum potential thesis. Applied Measurement in Education, 13, 99–117. 10.1207/s15324818ame1301_5 [DOI] [Google Scholar]