Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2010 Jun 1.
Published in final edited form as: Psychol Aging. 2009 Jun;24(2):462–475. doi: 10.1037/a0014417

Metacognitive Influences on Study Time Allocation in an Associative Recognition Task: An Analysis of Adult Age Differences

Jarrod C Hines 1, Dayna R Touron 2, Christopher Hertzog 3
PMCID: PMC2690717  NIHMSID: NIHMS94370  PMID: 19485662

Abstract

The current study evaluated a metacognitive account of study time allocation, which argues that metacognitive monitoring of recognition test accuracy and latency influences subsequent strategic control and regulation. We examined judgments of learning (JOLs), recognition test confidence judgments (CJs), and subjective response time (RT) judgments by younger and older adults in an associative recognition task involving two study-test phases, with self-paced study in phase 2. Multilevel regression analyses assessed the degree to which age and metacognitive variables predicted phase 2 study time independent of actual test accuracy and RT. Outcomes supported the metacognitive account – JOLs and CJs predicted study time independent of recognition accuracy. For older adults with errant RT judgments, subjective retrieval fluency influenced response confidence as well as (mediated through confidence) subsequent study time allocation. Older adults studied items longer which had been assigned lower CJs, suggesting no age deficit in using memory monitoring to control learning.


An individual attempting to learn new information in preparation for a test must decide how much effort to spend studying and re-studying that information. Self-regulation of study behavior involves selectively choosing information to study and choosing how – and for how long – to study that information (e.g., Winne & Perry, 2000). Research on metacognition and self-regulated learning examines how individuals use monitoring and control mechanisms to optimize learning (e.g., Nelson, 1993, 1996). A general premise is that learners' selection of items for study or allocation of study time is influenced by their current state of learning, goals for future learning, and beliefs about themselves as learners (Cull & Zechmeister, 1994; Dunlosky & Hertzog, 1998; Metcalfe & Kornell, 2005; Thiede & Dunlosky, 1999). For example, individuals may choose to devote more study time to information they have not yet learned (e.g., Nelson & Narens, 1990). Typically, rates of learning are fastest when individuals devote more effort to studying items they have not yet mastered.

This study evaluates whether there are age differences in self-regulation of learning based on metacognitive monitoring. Although previous research indicates that older adults are capable of accurate encoding and retrieval monitoring (see Hertzog & Hultsch, 2000), the limited evidence available suggests older adults may be less likely to use their accurate monitoring to effectively regulate learning and test performance. In self-paced testing, older adults do not judge readiness to recall as accurately as younger adults (Murphy, Sanders, Gabrieshki, & Schmitt, 1987; Riggs, Lachman, & Wingfield, 1995), in part because they do not spontaneously use self-testing strategies to monitor recall readiness (but see Bottirolli et al, 2008). Older adults, like younger adults, are more likely to select items they have not yet learned for restudy when forced to choose a limited subset of items (Dunlosky & Hertzog, 1997). However, given the opportunity to restudy all items, older adults' self-paced study time may not be optimally allocated. Dunlosky and Connor (1997) found larger negative gamma correlations between recall and study time for younger adults than for older adults, suggesting that younger adults devote more time and effort to items they have not yet learned, relative to older adults. The difference might arise because older adults do not use the more optimal control strategy of differential allocation of study time to unlearned items, or because they invest more time on maintenance rehearsal of already learned items, fearful that otherwise these items will be forgotten. In either case, a simultaneous evaluation of metacognitive monitoring and control over study is needed to insure that any age differences that are observed are consistent with the hypothesis of spared monitoring but impaired control (Hertzog & Hultsch, 2000). Dunlosky and Connor (1997) supported the spared monitoring, impaired control hypothesis, because they found no age differences in the accuracy of delayed judgments of learning, but age differences in the correlations of delayed judgments of learning with subsequent study time allocation.

The present study allows for a more thorough examination of age differences in monitoring mechanisms that are relevant to self-regulated learning because it explicitly separates two classes of monitoring that are potentially relevant: (a) monitoring effectiveness of study and (b) monitoring retrieval at the time of test. In order to frame the problem in terms of which types of monitoring could be related to study time allocation, we turn next to a discussion of the relevant monitoring mechanisms.

Monitoring Study and Test Performance

A metacognitive perspective argues that monitoring at different stages of learning can be used to for self-regulatory control (e.g., Nelson & Narens, 1990). Judgments of learning (JOLs) -- confidence in future recall after information has been studied -- correlate with later memory performance and with study time allocation in tasks like paired associate learning (e.g., Nelson, 1993; Dunlosky & Connor, 1997). JOLs made immediately after each item is studied – but prior to any memory test for that item – assess monitoring of encoding. JOLs made with a delay after study, but still prior to any memory test, measure monitoring of accessibility of target information when cued (Dunlosky & Nelson, 1994; Nelson & Dunlosky, 1991; Weaver & Kelemen, 2003). Delayed JOLs are typically more accurate in predicting memory performance, but immediate JOLs may reflect aspects of monitoring encoding that persist over learning trials and that influence study time.

Much of the experimental literature on monitoring and study time allocation has evaluated JOLs (e.g., Koriat, May'an, & Nussinson, 2006; Metcalfe & Finn, 2008). Ironically, JOLs may not capture the most important aspect of monitoring relevant to study time allocation. Although delayed JOLs predict subsequent study behaviors, including item selection for re-study and study time allocation (e.g, Dunlosky & Connor, 1997; Nelson, Dunlosky, Graf, & Narens, 1994), correlations are typically stronger between study behavior and memory performance itself (Dunlosky & Connor, 1997).

If actual memory performance is the best predictor of study time allocation, then metacognitive monitoring of cognitive states during test, including both target accessibility and accuracy of memory test performance may be the more important sources of feedback for self-regulatory control, including study time allocation. Indeed, individuals discover which items are well learned versus unlearned principally on the basis of test experience (e.g., Dunlosky & Hertzog, 2000; Finn & Metcalfe, 2007; Hertzog, Dixon, & Hultsch, 1990; Hertzog, Dunlosky, & Price, in press; Koriat & Bjork, 2006). Confidence judgments (CJs) – rated confidence in the accuracy of responses following memory tests – are proximal indicators of performance monitoring (Dunlosky & Hertzog, 2000; Higham, 2002). One could even argue that the relationship of JOLs to study time could be a causally spurious correlation created by the fact that both JOLs and study time allocation correlate with memory performance monitoring (as measured by CJs) and memory (but see Metcalfe and Finn, 2008).

The potential importance of performance monitoring, as measured by CJs, leaves open the possibility that age differences in study time allocation are due to age-related deficits in the accuracy of performance monitoring rather than the monitoring mechanisms captured by JOLs. Older adults are prone to high-confidence errors in some recognition memory tasks (e.g., Dodson, Bawa, Krueger, 2007; Kelley & Sahakyan, 2003), and such monitoring failures could contribute to suboptimal study time allocation.

In this study, we applied a metacognitive perspective to study time allocation by evaluating the influence of JOLs and CJs on strategic control in an associative recognition task. In general, a metacognitive perspective posits the following: if a metacognitive variable (e.g., CJs) measures monitoring that is potentially relevant to control of study, and if that monitoring influences subsequent study behavior, then the metacognitive variable should predict study behaviors even when statistically controlling on the objective performance that was monitored (e.g., recognition accuracy). Despite the relevance of performance monitoring for self-regulated study, the experimental literature has focused exclusively on JOLs, not CJs in evaluating study time allocation. This could be due in part to the fact that CJs are highly correlated with cued recall accuracy (e.g., Dunlosky & Hertzog, 2000), the memory task typically employed in study time allocation studies. The current study employed a recognition memory task for paired associate items. Recognition CJs are not as highly correlated with successful recognition as they are with recall (e.g., Benjamin, 2003; Kelley & Sahakyan, 2003), so we were more likely to discern unique associations of CJs on subsequent study allocation, independent of memory performance itself.

A Model for Study Time Allocation

Figure 1 contains a model for self-paced study time allocation as examined in this experiment. The first study opportunity (phase 1) was experimenter-paced and followed by an associative recognition memory test. Two aspects of actual memory performance, the accuracy and latency of phase 1 recognition test responses (Figure 1 performance variables: Phase 1 Recognition Accuracy and Phase 1 Recognition RT), were sources of evidence, or cues (Koriat, 1997), that could be monitored to guide subsequent self-paced study time allocation (Phase 2 Study Time Allocation). Figure 1 also shows three metacognitive variables that could guide study time allocation in phase 2: Phase 1 Judgments of Learning, Phase 1 Confidence Judgments, and Phase 1 Response Time Estimates. As noted above, phase 1 JOLs capture learning confidence during phase 1 study, whereas phase 1 CJs capture performance monitoring during the recognition memory test in phase 1. Given that a metacognitive account of study time allocation assumes an influence of item monitoring independent of item difficulty (as reflected in memory performance), we hypothesized that phase 1 JOLs and CJs would predict phase 2 study time allocation independent of phase 1 recognition performance. An open question was whether JOLs would predict phase 2 study time allocation independently of phase 1 CJs. If study time allocation was driven only by performance monitoring, then JOLs would correlate with study time allocation, but this effect would be eliminated when controlling on CJs.

Figure 1.

Figure 1

Flow diagram demonstrating precursors of phase 2 study time.

Whereas CJs reflect monitoring of recognition memory accuracy, memory test RTs can also be diagnostic of the degree to which an item has been learned. Recognition memory RTs reflect accessibility of information from episodic memory. Fast, correct recognition responses are also associated with higher confidence (CJs) about memory accuracy (e.g., Benjamin & Bjork, 2000; Dosher, 1982). Recognition memory RTs decrease as a function of repeated testing and enhanced learning (e.g., Touron & Hertzog, 2004), indicating that retrieval fluency is a valid cue for the probability of subsequent remembering.

Given the relevance of retrieval fluency as a metacognitive cue, we sought to directly measure monitoring of phase 1 recognition RTs. Monitoring of performance latency, or subjective (i.e., perceived) fluency, reflects how quickly an individual believes they evaluated the recognition test item. It can be measured by requesting explicit RT estimates after each recognition trial. If subjective fluency is a metacognitive cue that guides strategic control, then subjective fluency should influence phase 1 CJs and phase 2 study time allocation independent of actual fluency, as measured by actual RTs. The next section of the Introduction further examines subjective fluency as a metacognitive variable.

Subjective Fluency as a Metacognitive Cue

Memory researchers have shown that illusions of processing fluency can be misattributed to memory for prior events, which has been labeled a fluency heuristic (e.g., Kelley & Jacoby, 1998; Whittlesea & Leboe, 2003). Such research has utilized various experimental manipulations to influence perceptions of fluency and examined effects on perceptions of familiarity, but has not explicitly asked individuals to estimate the latency of their responses in order to measure subjective fluency. There is a large literature on monitoring temporal duration (e.g., Block & Zakay, 1997; Craik & Hay, 1999). However, metacognitive research has typically measured subjective confidence in memory accuracy but not subjective experiences of processing fluency. Instead, measures of actual fluency (i.e., actual rather than perceived RT) have been studied with respect to their influence on metacognitive judgments (e.g., Hertzog, Dunlosky, Robinson, & Kidder, 2003; Koriat & Ma'ayan, 2005).

Explicit measurement of subjective fluency is important for evaluating metacognition and cognitive control because actual and subjective fluency are separable constructs and behave in empirically distinct ways. Robinson, Hertzog, and Dunlosky (2006) found that estimates of encoding time (a type of subjective fluency) were more correlated with JOLs than actual encoding time (actual fluency). Actual fluency of encoding and retrieval influences memory judgments independent of its influences on memory performance (e.g., Benjamin, Bjork, & Schwartz, 1998; Hertzog et al., 2003; Koriat & Ma'ayan, 2005), consistent with a metacognitive account of fluency effects. Older adults, like younger adults, show this effect (Robinson et al., 2006). Given that subjective encoding fluency influences JOLs independent of actual encoding RTs, we hypothesized that subjective fluency of recognition test responses would predict both recognition CJs and the allocation of further study time independent of actual recognition RTs. It may be the case that what is perceived to have occurred (subjective RT) is of greater importance than what actually occurred (i.e., actual RT) in terms of allocating future study time.

Inaccurate perceptions of fluency, if used to assess learning, might adversely impact strategic control. For example, an inaccurate perception of a slow recognition test response may lead a person to study an item longer than warranted based on the actual retrieval latency. Age differences in the accuracy of RT monitoring therefore represent another, as-yet-untested potential source of age differences in study time allocation. Older adults are less accurate than younger adults when monitoring latency in a variety of experimental tasks (e.g., Block, Zakay, & Hancock, 1998; Craik & Hay, 1999; Hertzog, Touron, & Hines, 2007). Particularly germane to this study, older adults are less accurate in estimates of retrieval RTs in associative recognition tasks compared to younger adults (Hertzog et al., 2007). However, Hertzog et al. (2007) showed that these deficits can be at least partially ameloriated by explicit RT feedback and training.

We therefore manipulated the accuracy of subjective fluency in this experiment by providing actual RT feedback on odd trials, alternating with trials that requested RT estimates. The advantage of this procedure is that we expected RT feedback to influence subjective fluency but not actual RTs, allowing us to consider the contribution of RT estimate accuracy for subsequent monitoring and control. Note in our Figure 1 model that feedback directly influences only subjective RT, not downstream variables such as CJs. The hypothesis of age differences in study time regulation can be tested by evaluating interactions of age with JOLs, CJs, and RT estimates, on subsequent study time allocation, as we discuss in more detail shortly.

Intact and Rearranged Pairings in Associative Recognition

The associative recognition test we employed required participants to discriminate between intact pairs (same pairings as originally studied) and rearranged pairs (words from originally studied items are recombined into new pairings). This approach prevents correct responding solely on the basis of recognizing words as old, without recovering the studied associations. Cohn and Moscovitch (2007) showed that strategies for correct rejection of rearranged lures are more complex because one can recall both or only one of the original associations and use that information to reject the rearranged lure. Moreover, age differences in associative recognition appear to be driven largely by a higher likelihood of false positives on rearranged items (Cohn, Emrich, & Moscovitch, 2008; Healy, Light, & Chung, 2005), possibly implicating age differences in retrieval-based inferences at test. For this reason, CJs for rearranged pairs may not reflect performance monitoring for both items involved in the rearrangement. A test of metacognitive control in the associative recognition task is therefore best assessed by examining relations of CJs for intact pairs to subsequent study time for those pairs. More broadly, it was important to evaluate whether the type of recognition test item (intact vs. rearranged) influenced metacognitive monitoring accuracy and subsequent study time allocation.

Summary of Major Hypotheses

The current research examined a metacognitive account of study time allocation in a two-phase associative recognition task, evaluating the influence of cues such as JOLs, CJs and subjective test fluency. A key feature of the design is that there are multiple, correlated cues that potentially influence subsequent study time allocation. In order to evaluate the more fundamental mediational hypotheses stated above, the typical approach in the experimental literature – calculation of within-person correlations of a single metacognitive variable (e.g., JOLs) with study time is inadequate for isolating influences of one cue, controlling on others. Instead, we used multilevel regression analyses (Singer, 1998; Snijders & Bosker, 1999), which allow for the simultaneous evaluation of multiple variables' influence on measures of interest (Hoffman & Rovine, 2007). Our analytic approach is described more fully in the Results section.

We anticipated that these multilevel regressions would reveal that metacognitive judgments (CJs and JOLs) account for independent variance in strategic control above and beyond objective performance measures. We predicted stronger relationships of CJs to study time allocation than of JOLs given the likely importance of performance monitoring for governing later study behavior. Furthermore, we expected that subjective fluency would impact metacognitive judgments and study allocation beyond any effects of actual fluency. Finally, we examined age differences in predictive relationships to address the possibility that older adults do not effectively utilize monitoring for the self-regulation of study behavior. In particular, subjective RT effects should be more robust for older adults because they are expected to display a greater discrepancy between actual and estimated recognition test RT—and this is especially true for those who do not receive objective RT feedback to aid in calibrating their RT estimates.

Method

Design

The experiment was a 2 (Age: Young, Old) × 2 (Response Time Feedback: Given, Withheld) × 2 (Study-Test Phase: 1, 2) mixed design, with age and feedback as between-subjects independent variables and study-test phase as the within-subjects variable.

Participants

Adults between the ages of 18 and 25 years were classified as young, while our older group consisted of adults between the ages of 60 and 75. Younger adult participants (n = 39, Mage = 21.8) were undergraduates from Appalachian State University who received course credit for their participation. Older adults (n = 35, Mage = 66.8) were recruited from the nearby community and were paid for their participation. Approximately half of the participants in each age group were tested within each feedback condition. All participants were pre-screened for basic health issues that could impede participation, and were required to have good corrected visual acuity (20/50 or better). No participants were removed from our analyses as a result of these exclusions.

Materials and Procedure

Informed consent was established prior to beginning the experimental session. Participants then completed a brief personal data questionnaire, the Shipley Institute of Living vocabulary test, and the Digit Symbol subtest of the Wechsler Adult Intelligence Scale. Data pertaining to participant characteristics are presented in Table 1.

Table 1.

Means (and standard deviations) of participant characteristics

Measure Yngfeedback Yngno feedback Oldfeedback Oldno feedback
Age (years)1 22.2 (5.4) 21.3 (2.5) 66.7 (3.9) 67.0 (3.8)
Number of Medications1 1.0 (1.3) 1.4 (1.4) 2.4 (1.4) 2.3 (1.8)
Vocabulary1 29.5 (3.7) 29.2 (4.0) 35.2 (4.0) 35.2 (2.7)
Digit Symbol1, 3 65.6 (8.5) 69.1 (11.4) 52.1 (12.0) 46.3 (11.5)
Digit Symbol Memory1 7.0 (2.0) 7.7 (1.5) 5.7 (2.5) 5.7 (2.2)

Note. Number of Medications = self- reported number of daily medications taken. Vocabulary = number correct out of 40 on the Shipley Vocabulary Test (Zachary, 1986). Digit Symbol = WAIS Digit-symbol subtest (Wechsler, 1981). Digit Symbol Memory = symbol recall memory following the WAIS Digit-symbol subtest (Wechsler, 1981).

1

Age main effect: p < .05

2

Feedback main effect: p < .05

3

Age × Feedback interaction: p < .05

The associative recognition task was programmed in Visual Basic 6.0. Stimuli were presented using a 15-point Arial font on a 15-inch LCD monitor with a resolution of 1024 × 768. Participants were seated at a height and distance that optimized their screen viewing and comfort. Self-paced instructions preceded each portion of the task.

The computer task consisted of two consecutive study-test phases, which were identical in format and stimuli save for one key difference: study was experimenter-paced for phase 1 (word-pairs were presented for 4 seconds for younger adults or 6 seconds for older adults) and self-paced for phase 2 (participants terminated item-level study by pressing the spacebar). The computer task included the following sequence: (a) experimenter-paced study 1, (b) test 1, (c) self-paced study 2, and (d) test 2.

During each of the 2 study phases, 60 word-pairs (e.g., IVY-BIRD) were presented individually in a randomized order. These word-pairs were constructed from 120 semantically unrelated nouns, and all study items remained unchanged between study phases 1 and 2. After studying each item, participants gave a JOL rating their ability to remember the item approximately 10 minutes later on a continuous scale from 0% to 100% confidence. After viewing the final study item in each phase, participants began self-paced testing for that phase wherein, due to a programming error, 31 (instead of an even 30, or half) of the test stimuli were intact study items (e.g., an item presented as IVY-BIRD during study would be presented as IVY- BIRD during test), and the other 29 of the test items were rearranged study items (e.g., IVY-BIRD and BARREL-STAR might be presented at test as IVY-STAR and BARREL-BIRD). There was no overlap between the 31 matched and 29 unmatched word pairs, and items were selected randomly to be presented either intact or rearranged. The foils were created by merely swapping the second word in each word-pair foil with the second word from another foil. Test items were generated randomly prior to the test portion of each phase. Word-pair constituents were always presented in capital letters and always retained their position on the left or right side of the word-pair. Prior to the presentation of each test item, a “+” was presented on the screen to orient participants to the location of the upcoming word-pair. Upon the presentation of the stimulus, participants were instructed to press one of two keys indicating whether or not that word-pair had been studied previously.

Following odd-numbered trials, participants in the RT feedback condition were provided with their actual RT (in seconds) for the preceding trial. Participants in the no-feedback condition received no additional information after each yes/no response but were presented with a blank screen for a time interval of the same length as that of the feedback screen (250ms). Participants were given no rationale for why we asked them to estimate their trial-level recognition RTs; they were merely asked to supply their best guesses. While this measure may be less intuitive than a JOL or CJ, there was no evidence from our sample that individuals misunderstood what we were asking them to supply to us. Following even-numbered trials, participants estimated their recognition RT by choosing 1 of 9 RT estimate categories, which divided their potential responses into half-second intervals beginning with “less than 0.5 second” and ending with “greater than 4.0 seconds.” Next, participants provided a CJ rating their confidence in the accuracy of the previous recognition trial on a continuous scale from 0% to 100% confidence, by entering any value within in that range. After all items were tested, participants were offered a short break prior to the beginning of the next study-test phase.

Results

We examined the following measures: recognition memory accuracy, recognition memory RT, RT estimates, JOLs and CJs, and phase 2 self-paced study time. First, we present analyses of age differences in basic variables, including Goodman-Kruskal gamma correlations for resolution (relative accuracy) of metacognitive judgments. This approach is common in the metacognitive literature, thereby facilitating comparisons to other studies. In particular, we report aggregated statistics (such as gamma correlations or median RTs) that are absent in multilevel regression models that utilize item-level data directly. Second, we present the relationships of multiple variables on CJs and study time allocation via multilevel regression models.

General Linear Model Analyses

All variables were analyzed in a general linear model (GLM) with age group (Young, Old) and RT feedback condition (Given, Withheld) as between-subjects factors, and study-test phase (1,2) and Item Type (Intact, Rearranged) as within-subjects factors.

Recognition Memory Accuracy

Younger and older adults had similar levels of recognition accuracy (see Table 2). Interestingly, intact test items produced lower accuracy than rearranged items, F(1, 70) = 4.00, p < .05, but this difference was small and only reliable during phase 1, F(1, 70) = 8.17, p < .01. Both age groups' recognition accuracy improved at phase 2, F(1, 70) = 128.19, p < .001. There was no age difference in learning over phases, F(1, 70) = 1.57, p = .22. Thus, varying encoding time for older and younger adults was sufficient to equate them on recognition memory, a useful feature for testing hypotheses about age differences in monitoring accuracy and study allocation.

Table 2.

Recognition memory accuracy by age, test phase, and word-pair type.

Younger Adults Older Adults
Phase 1 Phase 2 Phase 1 Phase 2
Word-pair Type M SD M SD M SD M SD
Intact 0.73 0.18 0.88 0.19 0.73 0.17 0.86 0.15
Rearranged 0.79 0.18 0.90 0.15 0.78 0.17 0.85 0.21

Recognition Memory RT

Analysis of RTs was restricted to correct recognition trials only, using median RTs to reduce the influence of extreme outliers. Younger adults responded faster than older adults to memory test items, F(1, 70) = 46.87, p < .001 (see Table 3). Intact items were responded to more quickly than were rearranged items, F(1, 70) = 82.61, p < .001, and this difference was greater for older adults, F(1, 70) = 18.79, p < .001. The magnitude of the difference was substantial, consistent with the argument that rearranged pairs invoke additional processes such as recall-to-reject or recall-to-accept that increase processing time (e.g., Cohn & Moscovitch, 2007). Reflecting improved learning, participants responded faster at phase 2, F(1, 70) = 96.90, p < .001, with older adult RTs decreasing more overall, F(1, 70) = 4.90, p < .05.

Table 3.

Recognition test response time by age, test phase, and word-pair type.

Younger Adults Older Adults
Phase 1 Phase 2 Phase 1 Phase 2
Word-pair Type M SD M SD M SD M SD
Intact 1754 585 1206 351 2932 961 2045 769
Rearranged 2015 718 1475 473 3666 1599 2808 1107

RT Estimate Accuracy

The absolute accuracy of RT estimates was calculated as the difference between estimated RT and actual RT (binned for analysis into 500ms intervals to match the scale used for RT estimates; see Table 4). Replicating previous work (Hertzog et al., 2007), older adults underestimated their RTs to a greater degree than younger adults, F(1, 70) = 22.36, p < .001, and those who received RT feedback estimated their RTs more accurately than did those who did not, F(1, 70) = 20.97, p < .001 (Mfb = −.85, SDfb = 1.55; Mno fb = −2.39, SDno fb = 2.03). More accurate RT estimates were given for intact items, F(1, 70) = 30.66, p < .001, and the difference in RT estimation accuracy between item types was greater for older adults, F(1, 70) = 7.50, p < .01. This may be a consequence of faster RTs for intact items; past research has shown that older adults estimate time less accurately as intervals increase (e.g., Craik & Hay, 1999; but see Hertzog et al., 2007). Response time monitoring accuracy improved in phase 2, F(1, 70) = 84.81, p < .001, with older adults showing greater improvement, F(1, 70) = 13.23, p < .01. This result, however, is qualified by older adults' greater potential for improvement; both groups may have comparably improved had they displayed similar initial monitoring accuracy. Regardless, a clear and substantial age difference in RT monitoring accuracy was found. While older adults' mean actual RT was approximately twice that of younger adults, their estimated RT displayed but a fraction of that difference.

Table 4.

Absolute accuracy of recognition test response time estimates by age, test phase, response time feedback condition, and word-pair type.

Younger Adults Older Adults
Phase 1 Phase 2 Phase 1 Phase 2
M SD M SD M SD M SD
Response Time
Feedback Given
−0.43 0.93 −0.05 1.01 −2.11 1.74 −0.87 1.13
Response Time
Feedback Withheld
−1.91 2.03 −1.12 1.76 −3.91 1.53 −2.66 1.54
Intact Word-pairs −0.95 1.64 −0.41 1.37 −2.71 1.78 −1.29 1.50
Rearranged Word-pairs −1.15 1.75 −0.67 1.56 −3.27 2.02 −2.20 1.98

Note: Negative values indicate underestimation of test response times. A mean value of zero would indicate perfect monitoring accuracy, or no difference between actual and estimated response latency.

Phase 2 Self-Paced Study Time

Older adults took reliably more time to study items, which we consider in detail with the multilevel regression models for this variable. A main effect of item type was also found whereby items that were rearranged at test 1 were studied longer during phase 2 compared to previously intact items, F(1, 67) = 10.95, p < .01 (see Table 5). Items incorrectly recognized at phase 1 were also studied longer than those which were correctly recognized, F(1, 70) = 33.75, p < .001. However: rearranged items were studied longer only if accurately recognized, F(1, 67) = 19.01, p < .001.

Table 5.

Phase 2 study time by age, phase 1 recognition test accuracy, and word-pair type.

Younger Adults Older Adults
Accurate Inaccurate Accurate Inaccurate
Word-pair Type M SD M SD M SD M SD
Intact 2044 1367 3866 2570 4512 2990 8096 5104
Rearranged 3410 2162 3629 2705 6809 4510 7746 5964

Metacognitive Judgments

No main effects of age were found for in magnitudes of JOLs or CJs (see Table 6). Mean judgments increased during phase 2 (JOLs: F(1, 70) = 96.50, p < .001; CJs: F(1, 70) = 46.36, p < .001), consistent with increases in recognition memory accuracy. Younger adults' JOLs increased more between phases than did those of older adults, F(1, 70) = 4.43, p < .05. Intact items were given higher recognition confidence ratings overall compared to rearranged items, F(1, 70) = 30.87, p < .001, but the difference was not reliable in phase 2, F(1, 70) = 6.44, p < .05.

Table 6.

Metacognitive memory measures by age, test phase, and word-pair type.

Judgments of Learning
Younger Adults Older Adults
Phase 1 Phase 2 Phase 1 Phase 2
Word-pair Type M SD M SD M SD M SD
Intact 29.32 21.89 63.85 29.68 35.00 23.26 54.60 27.61
Rearranged 31.68 24.45 64.15 31.01 31.64 22.19 54.74 26.20
Confidence Judgments
Younger Adults Older Adults
Phase 1 Phase 2 Phase 1 Phase 2
Word-pair Type M SD M SD M SD M SD
Intact 82.14 29.87 91.56 22.67 72.36 29.67 84.66 27.96
Rearranged 73.12 28.00 87.51 26.24 55.83 32.40 78.54 29.24

Due to Nelson (1984), Goodman-Kruskal gamma correlations have become the traditional method for assessing resolution or relative accuracy of metacognitive judgments, although this method of analysis is not always ideal.1 These correlations are reported in Table 7; the df are lower than in other analyses because gammas cannot be calculated for individuals with recognition accuracy performance at floor or ceiling. No age differences in resolution were found for either JOLs (F(1, 70) = 1.51, p > .25) or CJs (F(1, 70) = 1.68, p > .19). As hypothesized, JOLs correlated more weakly with test performance than CJs, but resolution in both cases was above chance (reliably greater than 0). Overall JOL resolution increased between phases, F(1, 70) = 6.72, p < .05. In addition, the relative accuracies of both JOLs and CJs (when collapsed across phase) were greater for intact items, F(1, 70) = 25.72, p < .001, and F(1, 70) = 13.61, p < .001, respectively.

Table 7.

Gamma correlations between metacognitive measures and phase 1 recognition test accuracy by age, test phase, and word-pair type.

Judgments of Learning
Younger Adults Older Adults
Phase 1 Phase 2 Phase 1 Phase 2
Word-pair Type M SD M SD M SD M SD
Intact 0.28 0.41 0.46 0.53 0.25 0.49 0.42 0.49
Rearranged −0.03 0.49 0.26 0.55 −0.04 0.51 0.02 0.68
Confidence Judgments
Younger Adults Older Adults
Phase 1 Phase 2 Phase 1 Phase 2
Word-pair Type M SD M SD M SD M SD
Intact 0.64 0.47 0.72 0.61 0.62 0.34 0.59 0.62
Rearranged 0.42 0.5 0.48 0.74 0.12 0.47 0.45 0.52

To evaluate whether metacognitive judgments correlated with study time allocation, gamma correlations were computed between our metacognitive memory measures at phase 1 and study time at phase 2. As expected, correlations were negative, with higher confidence associated with less study time at the next study opportunity. No age differences in the study time gamma correlation were found for either JOLs (F(1, 72) = .61, p > .4) or CJs (F(1, 72) = .39, p > .50), see Figure 2. The correlation of CJs with study time differed by item type, being dramatically lower for rearranged pairs, F(1, 72) = 62.38, p < .001. JOL-study time correlations were fairly low overall. The interaction between Age and Item Type was not reliable (.10 > p > .05).

Figure 2.

Figure 2

Gamma correlations between phase 2 study time and metacognitive measures of memory performance collected in phase 1, by age group and item type.

In summary, a common thread among GLM analyses was the distinction between intact and rearranged test trials. Because rearranging items impacted relationships between phase 1 accuracy and phase 1 CJs, as well as between these two variables and phase 2 study time, we concluded that monitoring of rearranged pairs was indeed complicated by the involvement of two items in each rearrangement, possibly due to different recognition strategies. Therefore, the multilevel regression analyses reported below utilized only test trials for intact word-pairs, which have a more definitive and straightforward interpretation as to how monitoring should affect control.

Multilevel Regression Analyses

Multilevel regression models allow for the simultaneous evaluation of multiple metacognitive variables relevant to hypotheses of statistical mediation (Hoffman & Rovine, 2007). The models were used to test metacognitive hypotheses of interest in the present study. For example, the hypothesis that subjective RTs completely mediate effects of actual RTs on CJs predicts that subjective RTs will significantly predict CJs controlling on actual RT, but that actual RTs will not predict CJs when controlling on subjective RTs. Incomplete or partial mediation would be revealed if both variables independently influence CJs, controlling on one another. The advantage of regression approaches is that correlated variables can be evaluated for their independent, direct and indirect influences on downstream variables.

Multilevel regression also allows for simultaneous consideration of within-person effects and between-person effects.2 For instance, it is possible to evaluate whether within-person (item) variation in CJs is associated with within-person variation in study time allocation (i.e., individuals allocate more study time to items they are not confident they know, and less time to items they are confident they know), as reflected in gamma correlations of these two variables. At the same time, one can also evaluate whether between-person differences in mean CJs are associated with between-person differences in average study time (i.e., individuals with overall lower confidence will allocate more overall study time compared to individuals with overall higher confidence). In principle, these influences can operate in different directions (as is often the case in speed-accuracy tradeoffs, where the tradeoff is a negative within-person correlation, but speed and accuracy are positively correlated between-persons). In such cases, the two types of influence can suppress each other if they are not separated (Snijders & Boskers, 1999). Note that within-person and between-person are standard terminology in multilevel regression, and are not necessarily analogous to within-subjects and between-subjects variability components in comparisons such as ANOVA.

We used multilevel regression models to evaluate predictors of phase 1 recognition test confidence and phase 2 study time. The relationships of interest can be seen in Figure 1. Items defined the first-level, characterized by the variables Phase 1 Recognition Accuracy, Phase 1 Recognition RT, Phase 1 Response Time Estimate, Phase 1 Judgment of Learning, Phase 1 Confidence Judgment, and Phase 2 Study Time Allocation. The second level was defined by participants, and included between-subjects variables of Feedback condition and Age group. SAS PROC MIXED (Littell, Milliken, Stroup, & Wolfinger, 2000) was used to estimate unstandardized regression equations linking the variables.

Our general approach was to include all relevant within-person and between-person variables in a regression analysis, and then trim nonsignificant effects to achieve a final regression model.3 We also routinely modeled random effects in intercepts and for test accuracy (and their correlation), reflecting between-subject differences in the dependent variable; we report reliable random effects only (p < .05).4 As discussed in the GLM section of the Results, test trials that included intact word-pairs were used exclusively in the regressions due to indications that the relationships between variables were less stable for trials that included rearranged word-pairs. The final regression results for the two critical equations involving prediction of phase 1 CJs and prediction of phase 2 study time are reported below. The reasons for the inclusion of these equations are as follows: (a) Age differences in study time allocation were the driving force behind the current investigation and (b) Phase 1 CJs were hypothesized to have a proximal and relatively large impact on phase 2 study time, so it was important to determine which variables relate to recognition test confidence.

Confidence judgments

For CJs, our initial regression equation included the variables of actual RT, estimated RT, recognition accuracy, JOLs, RT feedback, age, and selected interactions (feedback × estimated RT, age × JOL, age × actual RT, age × estimated RT, and age × recognition accuracy). After trimming nonsignificant effects, we found reliable influences of both within-person and between-person effects. Table 8 reports the regression estimates and significance tests for reliable predictors, with the estimates scaled in the metric of the original variables. Note that each effect is estimated while considering the effects of the other variables in the equation. Also note that age and RT feedback interactions, when reliable, are described with group-specific regression coefficients and t-values within the text.

Table 8.

Multilevel regression analysis predicting Phase 1 confidence judgments (intact word-pairs).

Effect Estimate SE df t
Intercept 64.48 5.20 68 12.413
Age (Younger) 9.57 5.64 68 1.70
Feedback Condition (Feedback Given) 0.58 6.11 68 0.10
Phase 1 Judgment of Learning
(Between Person)
0.50 .14 68 3.633
Phase 1 Recognition Accuracy
(Within Person)
20.36 3.04 1004 6.693
Phase 1 Response Time
(Within Person)
−0.003 .0005 1004 −6.123
Phase 1 Response Time (Within Person)
× Age (Younger)
−0.003 .001 1004 −2.361
Phase 1 Response Time Estimate
(Within Person)
−3.41 .82 1004 −4.163
Phase 1 Response Time Estimate
(Within Person)
× Feedback Condition (Given)
3.17 .97 1004 3.252
Phase 1 Response Time Estimate
(Between Person)
−7.05 2.78 68 −2.541
Phase 1 Response Time Estimate
(Between Person) × Age (Younger)
9.76 3.68 68 2.651

NOTE:

1

p < .05

2

p < .01

3

p < .001

We describe and evaluate significant predictors of CJs beginning at the top of the table. Given dummy variable coding on between-subjects factors of age and feedback, the intercept estimate corresponded to mean confidence of 64% in the older adults not given RT feedback. Age and feedback effects were not reliable, p > .05. An increase of 1% in a person's mean JOLs was associated with a .5% increase in test confidence, independent of any effect of actual recognition accuracy, though correct recognition responses were indeed rated higher in confidence than those which were not recognized correctly.

A reliable random effect of equation intercept was found, estimated variance = 567 (SE = 103, Z = 5.51, p < .001). We also detected reliable random effects for test accuracy (not shown in Table 8), indicating individual differences in the degree to which CJs were related to recognition memory performance, estimated variance = 213 (SE = 57, Z = 3.75, p < .001). Thus, individuals varied reliably in the resolution of their CJs. Given that participants were approximately 20 percent more confident in accurate compared to inaccurate responses, with a standard deviation of 26 for the random effect, it appears that many individuals had moderate to poor resolution of CJs with respect to recognition accuracy.5

Differences in RT were also related to test confidence. Actual RTs were scaled in ms. Exploring the obtained within-person Age × RT interaction further, we found that younger adults experienced a 6% drop in confidence per 1s increase in RT (β = .006, t = −4.93, p < .001), whereas older adults experienced a 3% drop in confidence per 1s increase in RT (β = .003, t = −5.77, p < .001). Younger adults' test confidence therefore appeared to display a greater sensitivity to items' actual RT. However, we cannot discount the possibility that this interaction is illusory, as it may be the case that younger and older adults were monitoring their RTs in a more relative than absolute sense. Because younger adult RTs were approximately half the magnitude of older adult RTs, a 6% drop per in confidence per second increase in younger RT could be equivalent to a 3% drop in confidence in older RT.

Subjective response time estimates were scaled in 500 ms bins. There was a reliable effect of Estimated RT (within subjects) on CJs, with faster estimated RTs associated with higher confidence. This effect was qualified by an interaction with RT Feedback condition. The effect of Estimated RT was not reliable for those who received RT feedback (β = −.72, t = −1.11, p > .25), but there was an average within-person decrease of about 2% confidence per RT estimate bin for those who did not (β = −2.43, t = −2.57, p < .05). Converting the binned estimates, those who did not receive feedback displayed a reliable 4% decrease in confidence per second increase in subjective fluency. Thus, feedback increased the correspondence of estimated RTs to actual RTs and thereby eliminated the effect of estimated RT on CJs. Absent feedback, subjective fluency and actual fluency were independently related to participants' confidence in their recognition memory responses at the item-level. Between-person differences in RT monitoring were also related to CJs. An increase of one RT estimate bin (i.e., .5 perceived seconds) was associated with a 4.5% increase in younger adult CJs (β = 4.52, t = 2.18, p < .05) and a 7.8% decrease in older adult CJs (β = −7.78, t = −2.27, p < .05). Each age group may have interpreted lower response fluency differently (i.e., a more careful versus a more difficult memory search).

To summarize, both actual performance and metacognitive factors displayed independent relationships with recognition test confidence during phase 1. Within-person, differences in actual RT were more related to younger adults' CJs, while differences in subjective or estimated RT were more related to older adults' CJs. For participants not given RT feedback, individual differences in subjective fluency were related to CJs, and could also directly or indirectly influence future study allocation.

Study time allocation

Table 9 reports the final regression model predicting phase 2 study time allocation. The intercept was scaled to older adults, who took on average about 6.2 seconds for self-paced study. We found reliable random effects for the equation intercept, estimated variance = 3704791 (SE = 770648, Z = 4.81, p < .001). Younger adults studied approximately 3 seconds less than older adults. Individuals studied items that were recognized correctly during phase 1 about 1.2 seconds less than items that were unrecognized.

Table 9.

Multilevel regression analysis predicting phase 2 study time allocation (intact word-pairs).

Effect Estimate SE df t
Intercept 6210.0 356.1 70 17.443
Age (Younger) −3158.1 488.7 70 −6.463
Phase 1 Recognition Accuracy
(Within Person)
−1237.4 309.9 993 −3.993
Phase 1 Judgment of Learning
(Within Person)
−17.3 5.9 993 −2.962
Phase 1 Judgment of Learning
(Between Person)
61.2 17.1 70 3.583
Phase 1 Judgment of Learning
(Between Person) × Age (Younger)
−90.6 23.5 70 −3.863
Phase 1 Confidence Judgment
(Within Person)
−24.4 7.3 993 −3.363
Phase 1 Confidence Judgment
(Within Person) × Age (Younger)
21.2 9.2 993 2.301
Phase 1 Response Time (Within Person) 0.3 0.1 993 4.103
Phase 1 Response Time Estimate
(Within Person)
233.0 120.2 993 1.94
Phase 1 Response Time Estimate
(Within Person) × Age (Younger)
−319.2 162.2 993 −1.971

NOTE:

1

p < .05

2

p < .01

3

p < .001

Both within-person and between-person effects of phase 1 study confidence (i.e., JOLs) were found, as were separate between-person effects by age group. Participants studied items 17 ms less during phase 2 for each point increase in phase 1 JOL; perceptions of initial item difficulty therefore persisted to phase 2. Figure 3, plots the age-related interactions of our metacognitive variables and phase 2 study time. A 1% increase in between-person JOL confidence was associated with a 29 ms decrease in study time for younger adults, but a 61 ms increase in study time for older adults. Whereas younger adults with overall higher JOLs predictably allocated less study time at phase 2, older adults with overall higher JOLs allocated more study time at phase 2. Perhaps older adults who were confident in their learning after phase 1 study engaged in more maintenance rehearsal to maintain their perceived high level of item learning, or perhaps these older adults were more sensitive to recognition difficulty at phase 1 test and consequently studied more at phase 2.

Figure 3.

Figure 3

Age interactions with phase 1 between-person judgments of learning, within-person confidence judgments, and within-person response time estimates on phase 2 study time allocation.

Also critical to the metacognitive hypothesis, there was a reliable within-person effect of CJs controlling on test accuracy, with higher CJs associated with shorter study time at phase 2, as predicted. This effect was also qualified by a significant age interaction (see Figure 3). For each 1% increase in phase 1 CJs, younger adults studied 3 ms less, while older adults studied 24 ms less. Items with lower confidence at Test 1 were indeed studied longer by older adults at phase 2. The magnitude of the slopes was reliably larger for older adults, a finding inconsistent with the hypothesis of an age-related deficiency in the utilization of monitoring.

A within-person increase of 1 ms in phase 1 recognition RT was associated with an increase of 3 ms in phase 2 study time. Participants therefore discriminated between items based upon phase 1 actual RT and allocated phase 2 study time accordingly. There was no reliable main effect for phase 1 within-person estimated RT, although there was a reliable interaction of this variable with age (see Figure 3). For each bin increase in estimated RT, younger adult study times decreased by 87 ms, while older adult study times increased by 233 ms. Neither coefficient was reliable separately in the two age groups, so we cannot conclude there is an effect of subjective fluency on study time allocation, independent of other metacognitive judgments. Given the robust effects of RT estimates on CJs (Table 8), it appears the principal source of an effect of RT monitoring accuracy on study time allocation was mediated through CJs. Note, however, that the effects of RT estimates on CJs were only reliable when RT estimates were inaccurate (i.e., in the condition with no RT feedback), and that the effects of CJs on study time were reliable only for older adults. For example, an older adult who erroneously perceived greater response fluency at phase 1 test may have been more confident in their item knowledge. The individual may have therefore studied this item less in phase 2 than they would have with accurate performance monitoring. In this way, RT monitoring accuracy could influence both phase 1 recognition confidence and (mediated by CJs) phase 2 study time.

To summarize, both actual and metacognitive factors exerted independent effects on phase 2 study time allocation. Phase 1 JOLs influenced phase 2 study independent of the influence noted on recognition confidence, with less study for items that had higher JOLs, and contrasting individual difference patterns by age group such that younger adults with high JOLs studied less, whereas older adults with high JOLs studied more. Recognition CJs also showed different effects for the two age groups, such that older adults (but not younger adults) studied items less if they had been confident in their recognition response. The actual RT for item recognition impacted item study time independent of the influence noted through CJs. Subjective fluency did not have a direct effect on study time allocation, independent of JOLs and CJs.

Discussion

These data provide clear evidence that study time allocation is related to metacognitive processes, independent of levels of performance. Both phase 1 JOLs and CJs predicted phase 2 study time allocation, controlling on phase 1 recognition memory performance. Although the relationship between memory performance and study time allocation could reflect effects of item difficulty that are independent of metacognitive control (Koriat et al., 2006), our outcomes for JOLs and CJs indicate that study time is also influenced by metacognitive factors. Specifically, when individuals have lower confidence in (a) the quality of their initial item learning and (b) the accuracy of their recognition response, they study that item longer at phase 2.

Older adults allocate study time during self-paced study based on metacognitive monitoring in a manner similar to young adults. Consistent with the hypothesis of spared monitoring, there were no reliable age differences in the resolution of JOLs or CJs, which were similarly correlated with recognition memory accuracy for both age groups. However, our results did not support the hypothesis of impaired utilization of monitoring. Older adults showed a robust effect of JOLs and CJs on phase 2 study time, and CJs actually had a greater influence on study time for older adults relative to younger adults. These results are inconsistent with a metacognitive control deficit, as exemplified by Dunlosky and Connor's (1997) findings regarding older adults' study time allocation. There are a number of differences in method between the studies, including the criterion memory test (cued recall in Dunlosky and Connor's study versus associative recognition in this study). One advantage of the present study is that our manipulation of presentation time – providing more study time for older adults in phase 1 -- was successful in equating younger and older adults in recognition memory test performance, making comparisons of relations between recognition accuracy, CJs, and study time allocation more interpretable. Dunlosky and Connor's experiment produced substantial age differences in recall performance at phase 1, consistent with the literature on slowed rates of learning in older populations (Kausler, 1994). Resolution of the discrepancy will require new experiments.

We note that the present results are consistent with other findings that older adults use a similar approach to selecting items for restudy based on past memory test outcomes (Dunlosky & Hertzog, 1997). It seems that age-related differences in metacognitive control of learning emerge under some conditions and not others. At present we do not understand the moderating variables that determine whether age differences in metacognitive control will arise.

The present study differs from earlier work in its focus on performance monitoring and CJs. CJs had a higher gamma correlation with study time allocation than JOLs, corresponding to the hypothesis that performance monitoring may be most relevant to future study time allocation. However, when all variables were in the multilevel regression model, the regression effects for CJs and JOLs were comparable. Immediate JOLs can reflect a number of aspects of encoding that are related to study time, including whether a mediator was formed and the fluency of encoding (Hertzog et al., 2003; Robinson et al., 2006). Thus, although the present results justify evaluating performance monitoring via CJs, they also implicate other metacognitive influences independent of CJs or recognition memory accuracy. On the other hand, the fact that recognition accuracy influences study time allocation independent of CJs and JOLs indicates that explicit metacognitive monitoring processes per se do not adequately capture all influences on study time (Koriat et al., 2006). The complex pattern of outcomes indicates the importance of considering multiple variables' influences in a simultaneous method, such as multilevel regression.

An interesting new finding in this study is the relationship of subjective RT estimates to CJs and study time allocation. Subjective fluency of recognition responses, as measured by phase 1 recognition RT estimates, reliably predicted CJs independent of actual RTs (actual fluency) when participants are provided no feedback about actual RT. This finding is interesting for two reasons. First, it is, to our knowledge, the first demonstration that subjective retrieval fluency has an impact on response confidence, and (mediated through confidence) on subsequent study time allocation. Second, it suggests that retrieval fluency effects that have been observed in the literature on memory and metamemory may be associated with conscious awareness and perhaps even explicit monitoring of retrieval fluency. Given that fluency is a construct that has had considerable impact on theorizing about metacognition (e.g., Benjamin et al., 1998; Hertzog et al., 2003; Koriat & Ma'ayan, 2005), the present result should focus attention on the benefits of explicitly measuring subjective fluency. For instance, an interesting question is whether aspects of fluency impact cognition and metacognition independent of any subjective awareness of fluency itself (e.g., Jameson, Narens, Goldfarb, & Nelson, 1994).

Another important contribution of the current study is the demonstration of an age-related reduction in the accuracy of estimated RTs during a recognition memory task (see also Hertzog et al, 2007). Both younger and older adults underestimated recognition test RT without corrective feedback, but the degree of underestimation was larger for older adults. This deficit was associated with a larger within-persons relationship between estimated RT and recognition CJs for younger adults, but a larger between-persons relationship between estimated RT and CJs for older adults. The within-person outcome suggests that younger adults' CJs across items are more affected by variability in RT across items, whereas in the older group, mean CJs are more influenced by mean estimated RTs. This latter result raises the possibility that illusory overconfidence by older adults in recognition tests is influenced by a global false impression of fast responding. Note that yes-no item recognition tests, which (a) are differentially influenced by fast RT for familiarity-based responses (Yonelinas & Jacoby, 1994) and (b) show greater age dependence on familiarity over recollection (e.g., Jennings & Jacoby, 1997; see Light, Prull, La Voie, & Healy, 2000), may be much more prone to illusory fluency effects than the type of associative recognition test we studied in this experiment.

Despite these age differences in subjective fluency, we found no age differences in the linkage of CJs to subsequent study time. This result parallels a finding of Robinson et al. (2006) regarding encoding fluency. In that study, older adults' subjective encoding fluency (measured by estimated encoding times) was less accurate than younger adults, but both age groups showed a similar effect of subjective encoding fluency on JOLs. In the present study, differential effects of retrieval fluency (as measured by estimated recognition RTs) on CJs did not result in differential effects of CJs on study time allocation for young and older adults. Such findings underscore the argument that one cannot merely rely on logical inference to assume that age differences in monitoring accuracy necessarily imply deficient metacognitive influences on control. An interesting question for future research is whether manipulations that influence the accuracy of CJs alter the correlation between memory performance and subsequent study time, by strengthening the performance-CJ relationship.

It would be useful to replicate and extend the present findings regarding subjective fluency and study time allocation to a modified associative recognition task. A limitation of the present study is that RT estimates were obtained on only half the trials (due to the inclusion of RT feedback). We successfully showed feedback effects on estimated RT, which supports the argument that this measure can be treated as subjective fluency that is distinct from actual RT. The feedback manipulation also allowed us to demonstrate that inaccurate RT estimates have a distinct impact on recognition confidence. However, it would be useful to evaluate subjective fluency in which estimates were obtained on all trials. Moreover, the binned RT estimate approach may have affected the magnitude of the relationship of estimated RT to CJs and study time allocation.

Another limitation of the present task was the fact that participants were tested on intact pairs for only half the items in phase 1. Rearranging study items at test had a substantial effect on CJs, and this effect appears to be of greater magnitude for older adults. Despite age equivalence in recognition accuracy, gamma correlations revealed that older adults were less confident in their recognition of word-pairs that were rearranged at test. Additionally, phase 1 CJs showed no correlation with phase 2 study time for rearranged items. This outcome can be understood in part because successful rejection of rearranged pairs can be made on the basis of recalling either of the two original pairings rather than both of them (e.g., Cohn & Moscovitch, 2007). Hence, one could be confident in rejecting a rearranged pair but have low confidence that one has learned one of the two constituent pairs involved in the rearranged test probe. Such an effect would dilute CJ-study time correlations. New research would be needed to evaluate these and other reasons why CJs for rearranged pairs were less accurate and did not predict study time allocation.

In conclusion, this study shows that metacognitive states – as captured by JOLs, CJs and subjective recognition test fluency – influence study time allocation independently of actual aspects of recognition memory performance such as accuracy and latency. These phenomena support a model of metacognitive self-regulation positing that monitoring affects control (Nelson, 1996). Furthermore, the results indicate no age deficits in the utilization of metacognitive monitoring to control study time allocation. This result departs from other results in the literature, and raises the possibility that age differences in self-regulation may depend on specific features of experimental tasks and the processes they invoke.

Acknowledgments

This research was supported by a grant from the National Institute on Aging, one of the National Institutes of Health (R01 AG0248). We thank Matt Meier, Zach Speagle, and Elizabeth Swaim for assistance with data collection.

Footnotes

1

Despite the common usage of the gamma correlation, it should be noted that such correlations involving study time for rearranged items should be interpreted with caution. While it is only possible to correctly recognize an intact item by remembering the association from a study trial, it is possible to correctly reject a rearranged word-pair by remembering the matching word for either the cue or target in the rearranged pair. Without knowing the source of the correct rejection, it is impossible to calculate a proper gamma. More general issues exist as well: first, gamma correlations focus on bivariate relationships (e.g., CJs and accuracy) ignoring all other influences on both variables; second, the comparison of gamma correlations risks aggregation bias in both estimation and inference (Singer, 1998).

2

In the current study, we separated within-person and between-person influences by creating two orthogonal transformed predictors for each variable. First, within-person measures of phase 1 recognition accuracy, response time, estimated response time, confidence, and study time for each person were computed by subtracting each person's mean for each variable from that person's trial-level scores on each variable (thereby creating a difference score for each trial). Centering on individual-level means allowed examination of within-person variation on the dependent variable from within-person variation of its predictors. Second, each person's mean on the relevant variable was computed (e.g., mean CJs), and then a between-persons difference score was computed by subtracting the sample mean for that variable from the person's mean. Centering on group-level means captured individual differences in the average scores of each variable. Because these two transformed variables are orthogonal, both can be included simultaneously in a regression equation. For example, a reliable negative effect of response time, centered at the individual level, on CJs would indicate that people were more confident in the accuracy of items they recognized quickly, versus those they recognized more slowly. On the other hand, a reliable negative effect of group-mean-centered response time would indicate that people who were generally slower to respond were generally less confident.

3

It should be noted that an attempt was made to examine more directly the role of stimulus effects on the relationships presented in the above multilevel regression analyses. This was accomplished by adding Stimulus and many different possible interaction terms, such as Stimulus × Age and Stimulus × Age × Phase 1 Test Accuracy, as independent variables in each equation. No reliable main effects or interactions involving stimulus were found for phase 1 test confidence, but a main effect of stimulus was found for phase 2 study time. Upon closer examination, the latter effect was limited to a few word-pairs which appeared to be less concrete than the others (e.g., EXERCISE-BANK and MANAGER-BLOCK), and the result was essentially that less concrete items were studied longer at phase 2. However, because stimulus effects were limited to so few items and failed to substantively impact the relationships between other variables in the equations presented above, we chose to exclude stimulus effects from the above-presented equations.

4

Individual differences in predictor effects are captured in estimates of random effects in slopes. Readers may wish to think of effects labeled as random in our regression models as having varying intercepts or varying slopes (see Gelman and Hill, 2007).

5

Assuming a normal distribution, 68% of the distribution of regression coefficients lay between approximately −6 and 46; participants therefore varied considerably in the degree to which test accuracy related to confidence reports.

Publisher's Disclaimer: The following manuscript is the final accepted manuscript. It has not been subjected to the final copyediting, fact-checking, and proofreading required for formal publication. It is not the definitive, publisher-authenticated version. The American Psychological Association and its Council of Editors disclaim any responsibility or liabilities for errors or omissions of this manuscript version, any version derived from this manuscript by NIH, or other third parties. The published version is available at http://www.apa.org/journals/pag/

Contributor Information

Jarrod C. Hines, Department of Psychology, Georgia Institute of Technology, Appalachian State University

Dayna R. Touron, Department of Psychology, Appalachian State University

Christopher Hertzog, School of Psychology, Georgia Institute of Technology..

References

  1. Benjamin AS. Predicting and postdicting the effects of word frequency on memory. Memory and Cognition. 2003;31:297–305. doi: 10.3758/bf03194388. [DOI] [PubMed] [Google Scholar]
  2. Benjamin AS, Bjork RA. On the relationship between recognition speed and accuracy for words rehearsed via rote versus elaborative rehearsal. Journal of Experimental Psychology: Learning, Memory, and Cognition. 2000;26:638–648. doi: 10.1037//0278-7393.26.3.638. [DOI] [PubMed] [Google Scholar]
  3. Benjamin AS, Bjork RA, Schwartz BL. The mismeasure of memory: When retrieval fluency is misleading as a metamnemonic index. Journal of Experimental Psychology: General. 1998;127:55–68. doi: 10.1037//0096-3445.127.1.55. [DOI] [PubMed] [Google Scholar]
  4. Block RA, Zakay D. Prospective and retrospective duration judgments: A meta-analytic review. Psychonomic Bulletin & Review. 1997;4:184–197. doi: 10.3758/BF03209393. [DOI] [PubMed] [Google Scholar]
  5. Block RA, Zakay D, Hancock PA. Human aging and duration judgments: A meta-analytic review. Psychology and Aging. 1998;13:584–596. doi: 10.1037//0882-7974.13.4.584. [DOI] [PubMed] [Google Scholar]
  6. Bottiroli S, Dunlosky J, Guerini K, Cavallini E, Hertzog C. Does strategy affordance moderate age-related deficits in strategy production? 2008 doi: 10.1080/13825585.2010.481356. Unpublished manuscript. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Cohn M, Moscovitch M. Dissociating measures of associative memory: Evidence and theoretical implications. Journal of Memory and Language. 2007;57:437–454. [Google Scholar]
  8. Cohn M, Emrich SM, Moscovitch M. Age-related deficits in associative memory: The influence of impaired strategic retrieval. Psychology and Aging. 2008;23:93–103. doi: 10.1037/0882-7974.23.1.93. [DOI] [PubMed] [Google Scholar]
  9. Craik FIM, Hay JF. Aging and judgments of duration: Effects of task complexity and method of estimation. Perception & Psychophysics. 1999;61:549–560. doi: 10.3758/bf03211972. [DOI] [PubMed] [Google Scholar]
  10. Cull WL, Zechmeister EB. The learning ability paradox in adult metamemory research: Where are the metamemory differences between good and poor learners? Memory and Cognition. 1994;22:249–257. doi: 10.3758/bf03208896. [DOI] [PubMed] [Google Scholar]
  11. Dodson CS, Bawa S, Krueger LE. Aging, metamemory, and high-confidence errors: A misrecollection account. Psychology and Aging. 2007;22:122–133. doi: 10.1037/0882-7974.22.1.122. [DOI] [PubMed] [Google Scholar]
  12. Dosher BA. Effect of sentence size and network distance on retrieval speed. Journal of Experimental Psychology: Learning, Memory, and Cognition. 1982;8:173–207. [Google Scholar]
  13. Dunlosky J, Connor LT. Age differences in the allocation of study time account for age differences in memory performance. Memory and Cognition. 1997;1997:691–700. doi: 10.3758/bf03211311. [DOI] [PubMed] [Google Scholar]
  14. Dunlosky J, Hertzog C. Older and younger adults use a functionally identical algorithm to select items for restudy during multi-trial learning. Journal of Gerontology: Psychological Sciences. 1997;52:178–186. doi: 10.1093/geronb/52b.4.p178. [DOI] [PubMed] [Google Scholar]
  15. Dunlosky J, Hertzog C. Aging and deficits in associative memory: What is the role of strategy production? Psychology and Aging. 1998;13:597–607. doi: 10.1037//0882-7974.13.4.597. [DOI] [PubMed] [Google Scholar]
  16. Dunlosky J, Hertzog C. Updating knowledge about encoding strategies: A componential analysis of learning about strategy effectiveness from task experience. Psychology and Aging. 2000;15:462–474. doi: 10.1037//0882-7974.15.3.462. [DOI] [PubMed] [Google Scholar]
  17. Dunlosky J, Nelson TO. Does the sensitivity of judgments of learning (JOLs) to the effects of various study activities depend on when JOLs occur? Journal of Memory and Language. 1994;33:545–565. [Google Scholar]
  18. Finn B, Metcalfe J. The role of memory for past test in the underconfidence with practice effect. Journal of Experimental Psychology: Learning, Memory, and Cognition. 2007;33:238–244. doi: 10.1037/0278-7393.33.1.238. [DOI] [PubMed] [Google Scholar]
  19. Finn B, Metcalfe J. Judgments of learning are influenced by memory for past test. Journal of Memory and Language. 2008;58:19–34. doi: 10.1016/j.jml.2007.03.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Gelman A, Hill J. Data Analysis using regression and multilevel/hierarchical models. Cambridge University Press; Cambridge, NY: 2007. [Google Scholar]
  21. Healy MR, Light LL, Chung C. Dual-process models of associative recognition in young and older adults: Evidence from receiver operating characteristics. Journal of Experimental Psychology: Learning, Memory, and Cognition. 2005;31:768–788. doi: 10.1037/0278-7393.31.4.768. [DOI] [PubMed] [Google Scholar]
  22. Hertzog C, Dixon RA, Hultsch DF. Relationships between metamemory, memory predictions, and memory task-performance in adults. Psychology and Aging. 1990;5:215–227. doi: 10.1037//0882-7974.5.2.215. [DOI] [PubMed] [Google Scholar]
  23. Hertzog C, Dunlosky J, Robinson AE, Kidder DP. Encoding fluency is used for judgments about learning. Journal of Experimental Psychology: Learning, Memory, and Cognition. 2003;29:22–34. doi: 10.1037//0278-7393.29.1.22. [DOI] [PubMed] [Google Scholar]
  24. Hertzog C, Hultsch DF. Metacognition in adulthood and old age. In: Salthouse TA, Craik FIM, editors. The handbook of aging and cognition. 2nd ed. Lawrence Erlbaum Associates; Mahwah, NJ: 2000. pp. 417–466. [Google Scholar]
  25. Hertzog C, Price J, Dunlosky J. How is knowledge generated about memory encoding strategy effectiveness? Learning and Individual Differences. doi: 10.1016/j.lindif.2007.12.002. in press. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Hertzog C, Touron DR, Hines JC. Does a time monitoring deficit influence older adults' delayed retrieval shift during skill acquisition? Psychology and Aging. 2007;22:607–624. doi: 10.1037/0882-7974.22.3.607. [DOI] [PubMed] [Google Scholar]
  27. Higham PA. Strong cues are not necessarily weak: Thomson and Tulving (1970) and the encoding specificity principle revisited. Memory & Cognition. 2002;30:67–80. doi: 10.3758/bf03195266. [DOI] [PubMed] [Google Scholar]
  28. Hoffman L, Rovine MJ. Multilevel models for the experimental psychologist: Foundations and illustrative examples. Behavior Research Methods. 2007;39:101–117. doi: 10.3758/bf03192848. [DOI] [PubMed] [Google Scholar]
  29. Jameson KA, Narens L, Goldfarb K, Nelson TO. The influence of subthreshold priming on metamemory and recall. Acta Psychologica. 1994;73:55–68. doi: 10.1016/0001-6918(90)90058-n. [DOI] [PubMed] [Google Scholar]
  30. Jennings JM, Jacoby LL. An opposition procedure for detecting age-related deficits in recollection: Telling effects of repetition. Psychology and Aging. 1997;12:352–361. doi: 10.1037//0882-7974.12.2.352. [DOI] [PubMed] [Google Scholar]
  31. Kausler DH. Learning and memory in normal aging. Academic Press; San Diego, CA, US: 1994. [Google Scholar]
  32. Kelly CM, Jacoby LL. Subjective reports and process dissociation: Fluency, knowing, and feeling. Acta Psychologica. 1998;98:127–140. [Google Scholar]
  33. Kelley CM, Sahakyan L. Memory, monitoring, and control in the attainment of memory accuracy. Journal of Memory and Language. 2003;48:704–721. [Google Scholar]
  34. Koriat A. Monitoring one's own knowledge during study: A cue-utilization approach to judgments of learning. Journal of Experimental Psychology: General. 1997;126:349–370. [Google Scholar]
  35. Koriat A, Bjork RA. Illusions of competence during study can be remedied by manipulations that enhance learner's sensitivity to retrieval conditions at test. Memory & Cognition. 2006;34:959–972. doi: 10.3758/bf03193244. [DOI] [PubMed] [Google Scholar]
  36. Koriat A, Ma'ayan H. The effects of encoding fluency and retrieval fluency on judgments of learning. Journal of Memory and Language. 2005;52:478–492. [Google Scholar]
  37. Koriat A, Ma'ayan H, Nussinson R. The intricate relationships between monitoring and control in metacognition: Lessons for the cause-and-effect relation between subjective experience and behavior. Journal of Experimental Psychology: General. 2006;135:36–69. doi: 10.1037/0096-3445.135.1.36. [DOI] [PubMed] [Google Scholar]
  38. Light LL, Prull MW, La Voie DJ, Healy MR. Dual-process theories of memory in old age. In: Perfect TJ, Maylor EA, editors. Models of cognitive aging. Oxford University Press; London: 2000. pp. 238–300. [Google Scholar]
  39. Littell RC, Milliken GA, Stroup WW, Wolfinger RD. SAS system for mixed models. 4th. Ed. SAS Institute, Inc.; Cary, NC: 2000. [Google Scholar]
  40. Mazzoni G, Cornoldi C. Strategies in study time allocation: Why is study time sometimes not effective? Journal of Experimental Psychology: General. 1993;122:47–60. [Google Scholar]
  41. Metcalfe J, Finn B. Evidence that judgments of learning are causally related to study choice. Psychonomic Bulletin & Review. 2008;15:174–179. doi: 10.3758/pbr.15.1.174. [DOI] [PubMed] [Google Scholar]
  42. Metcalfe J, Kornell N. A region of proximal learning model of study time allocation. Journal of Memory and Language. 2005;52:463–477. [Google Scholar]
  43. Nelson TO. A comparison of current measures of the accuracy of feeling-of-knowing predictions. Psychological Bulletin. 1984;84:93–116. [PubMed] [Google Scholar]
  44. Nelson TO. Judgments of learning and the allocation of study time. Journal of Experimental Psychology: General. 1993;122:269–273. doi: 10.1037//0096-3445.122.2.269. [DOI] [PubMed] [Google Scholar]
  45. Nelson TO. Consciousness and metacognition. American Psychologist. 1996;51:102–116. [Google Scholar]
  46. Nelson TO, Dunlosky J. When people's judgments of learning (JOLs) are extremely accurate at predicting subsequent recall: The “delayed-JOL effect.”. Psychological Science. 1991;2:267–270. [Google Scholar]
  47. Nelson T, Narens L. Metamemory: A theoretical framework and new findings. In: Bower GH, editor. The psychology of learning and motivation: Advances in research and theory. Vol. 26. Academic Press; 1990. pp. 501–508. [Google Scholar]
  48. Nelson TO, Dunlosky J, Graf A, Narens L. Utilization of metacognitive judgments in the allocation of study during multitrial learning. Psychological Science. 1994;5:207–213. [Google Scholar]
  49. Robinson AE, Hertzog C, Dunlosky J. Aging, encoding fluency, and metacognitive monitoring. Aging, Neuropsychology, & Cognition. 2006;13:458–478. doi: 10.1080/13825580600572983. [DOI] [PubMed] [Google Scholar]
  50. Singer JD. Using SAS proc mixed to fit multilevel models, hierarchical models, and individual growth models. Journal of Educational and Behavioral Statistics. 1998;24:323–355. [Google Scholar]
  51. Snijders TAB, Bosker RJ. Multilevel analysis: An introduction to basic and advanced multilevel modelling. Sage; London, U.K.: 1999. [Google Scholar]
  52. Thiede KW, Dunlosky J. Toward a general model of self-regulated study: An analysis of selection of items for study and self-paced study time. Journal of Experimental Psychology: Learning, Memory, and Cognition. 1999;25:1024–1037. [Google Scholar]
  53. Touron DR, Hertzog C. Strategy shift affordance and strategy choice in young and older adults. Memory and Cognition. 2004;32:298–310. doi: 10.3758/bf03196860. [DOI] [PubMed] [Google Scholar]
  54. Weaver CA, Kelemen WL. Processing similarity does not improve metamemory: Evidence against transfer-appropriate monitoring. Journal of Experimental Psychology: Learning, Memory, and Cognition. 2003;29:1058–1065. doi: 10.1037/0278-7393.29.6.1058. [DOI] [PubMed] [Google Scholar]
  55. Wechsler D. Wechsler Adult Intelligence Scale - Revised. Psychological Corporation; New York: 1981. [Google Scholar]
  56. Whittlesea BWA, Leboe JP. Two fluency heuristics (and how to tell them apart) Journal of Memory and Language. 2003;49:62–79. [Google Scholar]
  57. Winne PH, Perry NE. Measuring self-regulated learning. In: Boekarts M, Pintrich PR, Zeidner M, editors. Handbook of self-regulation. Academic Press; San Diego, CA: 2000. pp. 531–566. [Google Scholar]
  58. Yonelinas AP, Jacoby LL. Dissociations of processes in recognition memory: Effects of interference and response speed. Canadian Journal of Experimental Psychology. 1994;48:516–534. doi: 10.1037/1196-1961.48.4.516. [DOI] [PubMed] [Google Scholar]
  59. Zachary R. Shipley Institute of Living Scale Revised manual. Western Psychological Services; Los Angeles: 1986. [Google Scholar]

RESOURCES