Skip to main content
Cognitive Research: Principles and Implications logoLink to Cognitive Research: Principles and Implications
. 2024 Jul 12;9:46. doi: 10.1186/s41235-024-00572-8

Does using artificial intelligence assistance accelerate skill decay and hinder skill development without performers’ awareness?

Brooke N Macnamara 1,, Ibrahim Berber 1, M Cenk Çavuşoğlu 1, Elizabeth A Krupinski 2, Naren Nallapareddy 1, Noelle E Nelson 1, Philip J Smith 3, Amy L Wilson-Delfosse 1, Soumya Ray 1
PMCID: PMC11239631  PMID: 38992285

Abstract

Artificial intelligence in the workplace is becoming increasingly common. These tools are sometimes used to aid users in performing their task, for example, when an artificial intelligence tool assists a radiologist in their search for abnormalities in radiographic images. The use of artificial intelligence brings a wealth of benefits, such as increasing the efficiency and efficacy of performance. However, little research has been conducted to determine how the use of artificial intelligence assistants might affect the user’s cognitive skills. In this theoretical perspective, we discuss how artificial intelligence assistants might accelerate skill decay among experts and hinder skill acquisition among learners. Further, we discuss how AI assistants might also prevent experts and learners from recognizing these deleterious effects. We then discuss the types of questions: use-inspired basic cognitive researchers, applied researchers, and computer science researchers should seek to answer. We conclude that multidisciplinary research from use-inspired basic cognitive research, domain-specific applied research, and technical research (e.g., human factors research, computer science research) is needed to (a) understand these potential consequences, (b) design artificial intelligence systems to mitigate these impacts, and (c) develop training and use protocols to prevent negative impacts on users’ cognitive skills. Only by answering these questions from multidisciplinary perspectives can we harness the benefits of artificial intelligence in the workplace while preventing negative impacts on users’ cognitive skills.

Keywords: Artificial intelligence, Expertise, Skill acquisition, Skill decay

Introduction

Consider a radiologist examining a medical image for abnormalities. It is increasingly likely they will be aided by an artificial intelligence (AI) assistant (Hosny et al., 2018). The AI assistant may be a machine learning-based classifier that has been trained on a large number of images to, for example, classify abnormalities in breast images as benign or malignant. Given a new image, the AI highlights regions with presumed abnormalities or provides heatmaps, probabilities, graphs, or written descriptions. Such decision support systems are increasingly common in medicine and other areas (Magrabi et al., 2019).

As another example, the use of robots in surgical operations began over thirty years ago (see Gomes, 2011). Computer scientists are working to develop AI tools to assist the surgeon during operations. In the future, these AI assistants might gently nudge the surgeon to guide the surgical tools to the proper location, suggest the appropriate surgical tool for the task, and complete subcomponents of the procedure (Panesar et al., 2019).

In this theoretical perspective, we discuss the possibility that relying on AI assistants, including those yet to be developed, may have detrimental effects on human cognitive skills, such as expert judgment and decision making, integration of declarative knowledge, and problem solving. We use two subdisciplines within the field of medicine as examples: radiology, where artificial intelligence is currently being used in clinical practice, and robotic surgery, where artificial intelligence assistants are being developed to assist experts in the future. We first briefly describe cognitive processes involved in expertise and skill development. We then describe how the use of automated systems affects people’s cognitive skills and why consistent use of AI assistants, which differ from other forms of automation, may be detrimental to human skill learning and expertise. Specifically, we describe potential AI-induced skill decay and potential AI-induced skill development hindrance. Finally, we propose future directions that harness research from multiple perspectives are needed to understand these problems and develop solutions.

Learning and expertise

We acquire many skills over our lifetime to varying degrees of proficiency. Often, people initially rely on inefficient algorithms that are high in cognitive effort, such as children counting on their fingers to solve basic addition problems (Bajic & Rickard, 2011). With practice, we improve our performance, and for many skills, the amount of cognitive effort also dramatically decreases with practice (Fitts & Posner, 1967). Inefficient algorithms are replaced with more efficient algorithms or direct retrieval, such as adults directly accessing the answer to “2 + 2” from memory rather than walking through steps of a calculation (Ackerman & Woltz, 1994; Logan, 1988).

Depending on the domain, practice designed to improve domain-specific performance tends to explain between 0 and 20% of variance in performance (Macnamara et al., 2014). For simple tasks that tap automated processes (Schneider & Shiffrin, 1977; Shiffrin & Schneider, 1977), skill development emerges quickly with practice before asymptoting (Fitts & Posner, 1967). Individual differences in cognitive abilities are relevant initially, but less so after accumulating practice (Ackerman, 1986). In contrast, for complex tasks that tap controlled processes (as opposed to automated processes; Schneider & Shiffrin, 1977; Shiffrin & Schneider, 1977), skill development emerges more slowly and is highly dependent on individual differences (e.g., Ackerman, 1986). In tasks that demand controlled processes, practice has limited predictive power because people vary in their starting points, learning rates, and apogees, such that an hour of practice for one person does not confer the same benefit as an hour of practice for someone else (Macnamara et al., 2023).

Because many complex tasks continue to rely on cognitive resources (Ackerman, 1986), individual differences in cognitive abilities often predict skill acquisition rates and levels of ultimate expertise. For example, working memory capacity predicts music sight-reading performance (Meinz & Hambrick, 2010) and Texas Hold’em expertise (Meinz et al., 2012); fluid intelligence predicts piano skill acquisition (Burgoyne et al., 2019) and reading competence (Lechner et al., 2019); and cognitive processing speed predicts air traffic control skill levels (Ackerman, 1988) and simultaneous language interpreting expertise (Macnamara et al., 2011). Further, the predictive strength of both experiential and cognitive factors depends on characteristics of the task, such as the consistency of stimuli-response mappings (Ackerman, 1986; Macnamara & Frank, 2018; Macnamara et al., 2014), level of concurrence among subtasks (Frank & Macnamara, 2021; Hoffman et al., 2014), and rate of stimuli change (Hoffman et al., 2014; Macnamara & Frank, 2018; Macnamara et al., 2014). Given the demands of complex tasks on people’s cognitive resources, technological advances have been made to reduce cognitive load and human error during complex task performance.

The use of automation and artificial intelligence

Automation has been used for decades across multiple industries. Perhaps the most studied use of automation on human mental processes is the autopilot in aviation. Autopilots are designed to perform some of the human pilot’s tasks, such as landing the plane and keeping the wings level during the flight. Without autopilot, flying demands continuous attention from the human pilot. For longer flights, mental fatigue from the continuous attention became a safety threat. Thus, the use of automation has improved safety and flight performance (Federal Aviation Administration, 2013).

Despite these benefits, automation can also lead to bias and complacency, resulting in worse flight performance. When an automated system provides information to a user, this information can bias the user’s situation assessment (Smith et al., 1997), change their visual scanning patterns (Rezazade Mehrizi et al., 2023), and reduce how frequently they cross-check information (Manzey et al., 2012). In some cases, users are prone to favor information from automated systems over human expertise, even when the recommendations conflict and the automated suggestions are inaccurate (automation bias; Manzey et al., 2012; Mosier et al., 1997).

Other research indicates that using automation can lead to reliance on the automated system. Reliance can positively affect performance, such as when users appropriately trust the system to handle one part of the workload while the user directs additional attention to other components of the task. In contrast, reliance can negatively affect performance when users fail to evaluate the performance of automated systems and allow system errors, malfunctions, or anomalous conditions to go undetected (automation-induced complacency; Risko & Gilbert, 2016). Whether reliance is beneficial or detrimental depends on the accuracy of the automated system and how much trust the user has in the system. Quantitative syntheses of the literature suggest that the ‘crossover point’ from benefits to detriments, at least in high workload conditions, is around 70% accuracy (Parasuraman & Manzey, 2010; Wickens & Dixon, 2007). Yet, some users continued to rely on systems with low accuracy, such that their performance would have been better with no automation at all (Wickens & Dixon, 2007).

Automation bias and automation-induced complacency are not the same concerns as automation affecting a person’s skill. In other words, if the automation was removed, the user would no longer exhibit bias or automation-induced complacency. But what about the person’s skill? Some research has focused on how technological assistance might cause cognitive skill decay (Ebbatson et al., 2010; Kim et al., 2013; Kluge & Frank, 2014; Smith & Baumann, 2019; Volz et al., 2016), particularly in aviation (e.g., Casner et al., 2014; Ebbatson et al., 2010; Haslbeck & Hoerman, 2016). As an example, these studies raise a concern that the more time spent flying with automation support, the worse pilots’ performance was when this automation was unavailable or irrelevant (e.g., handling an aviation emergency, hand flying an approach). After a rise in near misses and other evidence that pilot’s manual skills were declining (Leahy & Fisher, 2023), the U.S. Federal Aviation Administration has recommended that pilots periodically use their manual skills for the majority of flights (Federal Aviation Administration, 2022).

Notably, automated systems, such as most autopilot systems, differ from artificial intelligence assistants. Automation refers to systems where the same input would repeatedly produce the same results. Automated system performance is governed by rules explicitly coded by developers. For example, airplane autopilots are automated control systems with a decision/feedback control loop based on known electromechanical characteristics of the airplane within normal parameters. When outside of normal parameters, such automation support will typically disengage and hand control to the pilots—they do not actively assist with expert decision making to the extent of AI.

In contrast, AI technologies are designed to mimic expert decision making by dynamically assessing problems, providing adaptive and novel solutions, and performing tasks that are cognitively demanding for humans (Walsh et al., 2019). These technologies come in multiple forms, from natural language processing to decision support systems (Bankins et al., 2023; Walsh et al., 2019). Here, we focus on a class of AI technologies—human-AI collaboration—often discussed in the future of work designed to aid humans in their jobs or skill development (Bankins et al., 2023; Brynjolfsson, 2023). More specifically, we refer to AI assistants as AI technologies (both current and future) that aid humans in problem solving, performance in cognitively demanding tasks, and in skill development. Examples include AI tools designed to aid radiologists during image analysis and AI tools designed to aid surgeons in their decisions and performance during robotic surgery operations. Current and future AI assistants go beyond automation to aid in high-level decision making that was previously the expert’s domain alone. To our awareness, no research has been conducted on AI assistants’ effects on the user’s skill.

Potential AI-induced skill decay

Research on skill decay suggests that after a period of disuse, skilled performance declines (Ackerman & Tatel, 2023; Arthur et al., 2013). A factor that influences this decay is the extent of cognitive demand in the task. More specifically, higher levels of skill decay are associated with tasks that require greater cognitive effort and fewer physical demands (i.e., controlled processes), as opposed to tasks that primarily rely on physical abilities with minimal cognitive processing (i.e., proceduralized skills) (Arthur et al., 1998; Wang et al., 2013).1

In addition to ceasing the task altogether, disuse can also refer to ceasing subcomponents of a task, such as when automation is used in place of the person applying their skills. As an illustration, Casner et al. (2014) tested the manual flying skills of pilots who were trained to fly manually, but then spent the majority of their careers flying with high automation (i.e., autopilot for much of takeoff, cruising, and landing). Procedural skills, such as scanning the instruments and manual control, were “rusty” but largely intact. In contrast, major declines in cognitive skills emerged, such as failures to maintain awareness of the airplane’s location, keep track of next steps, make planned changes along the route, and recognize and handle instrument systems failures.

To a much higher degree than automated systems, artificial intelligence assistants are designed to mimic cognitive skills—that is, they recognize patterns, reason about potential outcomes, and often guide the user to a specific action. For this reason, consistent and repeated engagement with an AI assistant is likely to lead to greater decrements in skill than engagement with automation systems. Frequent engagement with an AI assistant designed to take over some of the user’s cognitive processes means there are fewer opportunities to keep skills honed.

The potential for AI-induced skill decay is particularly relevant in high-risk industries, medical fields, aviation, and military operations, where maintaining high-level skills is essential for performance and safety (Day et al., 2013; Villado et al., 2013). Though some fields will flourish as AI improves in accuracy and takes over tasks, thereby reducing human error, other fields will have problems that AI cannot solve. For example, in medicine, new problems, such as a novel virus with unforeseen symptoms, can emerge at any time. Given that AI needs to be trained with data, human experts with keen insights are needed to solve novel problems. Such fields will benefit by maintaining human experts’ skills.

A potential complicating factor is that experts may be unaware of their skill decay. In cases where a person has ceased the task altogether, they will be aware that they have not used their skills and may be more likely to be aware of their skill decay. In contrast, AI-induced skill decay may operate outside the performer’s awareness because the disuse is only at the level of cognitive skill engagement, not with engagement with the task. As such, a surgeon using an AI assistant may believe their skills are still sharp because they have continued to perform operations at a high level (i.e., successful surgeries). They may not consider how well they would be able to perform without the AI assistant, such as selecting appropriate operative techniques, navigating complex anatomy (e.g., by identifying relevant anatomical landmarks and dissecting tissue to access target anatomy), or optimizing visualization (through changing laparoscopic camera angles and/or through dissection and retraction of tissue). Likewise, a radiologist using an AI assistant may believe they are highly capable of detecting and classifying anomalies in images, but their skills needed to discriminate between similar-appearing abnormalities, grade abnormalities severity properly, assess tissue density, measure tumor dimensions accurately, or detect subtle abnormalities may be less sharp than they believe. Both AI-induced skill decay and experts’ awareness of their skill decay are critical questions for use-inspired basic research, technical research, and applied research.

Potential AI-induced skill development hindrance

As AI assistants become increasingly prevalent, the role AI might have on skill development needs to be considered. AI-learning aids are designed to improve the rate of learner’s skill or knowledge development. We propose that frequent engagement with an AI assistant during skill development might hinder learning in some cases, depending on the ultimate goal for which the AI-learning aid was developed. Educational aids that are designed to personalize instruction for students with the goal of students independently performing the task are unlikely to hinder learning. In contrast, AI-learning aids designed to prepare trainees for work where AI assistants are used may focus on preparing the learner to work with an AI rather than focusing on developing learners’ cognitive skills independent of AI. For example, a radiologist trainee may not develop as keen visual detection skills or a surgical resident may not develop as robust spatial navigation skills as if they would have developed had they trained without assistance.

Here, it is important to distinguish between learning and performance. Suppose learners are randomly assigned to either learn a task with a high-performing AI assistant or without. We would expect the learners with the AI assistant to improve their performance rapidly and outperform the learners without AI. However, these performance gains may not reflect the learners’ gains in skill independent of the AI. That is, now suppose that after a period of learning that the AI is withheld, as may happen in the real world if the system is unavailable or fails. We might expect that the group who previously had access to the AI assistant to perform worse than those who never learned with the AI. In this case, those who learned with AI assistance might not have developed independent cognitive skills that the control group developed. Stated differently, we might expect to observe a pattern opposite of latent learning—high performance is observed in the AI-assisted group, but the limits on learning remain hidden until AI assistance is removed.

As with potential AI-induced skill decay, learners who have engaged with AI to assist in their skill development might be unaware of where their skills are lacking. In particular, AI assistants may promote illusions of understanding in learners, leading them to believe they have a greater understanding of the task than they actually do (Messeri & Crockett, 2024). These illusions of understanding may occur when learners believe they have a deeper understanding than they actually do (i.e., illusion of explanatory depth); when learners believe they are considering all possibilities rather than only those available through the AI assistant (i.e., illusion of exploratory breadth); and when learners believe that the AI assistant is objective, failing to consider the bias embedded in the AI tool from the developers and training data (i.e., illusion of objectivity) (Messeri & Crockett, 2024). As such, AI-induced skill development hindrance may not only limit the level of learning obtained but may change the nature of the understanding of the task.

Conclusions, recommendations, and future directions

AI tools are unquestionably beneficial in many respects (e.g., improved diagnostic accuracy, fewer human errors). However, a concern is that there may be downsides for the people using them. In particular, if the AI routinely aids performance at a high level, even well-trained experts may gradually lose their task-based cognitive skills, instead relying on the AI to render decisions. Additionally, it may be that trainees do not build their own cognitive skills at the task effectively when the AI serves as a learning aid. Further, because AI tools are likely to enhance performance and make the task feel easier, learners may be less able to judge the true status of their skills and experts may be unaware of their deteriorating skills, resulting in performance degradation when relying on the AI is not optimal.

Though the degradation of human skills may be of little concern in domains where reliance on AI can approach optimal performance, medicine will likely continue to need human experts. AI will have difficulty handling new viruses, unique injuries from accidents, or individual context and significant variations in a patient’s physiology. Thus, as AI assistants become increasingly common, their effects on human skills need to be considered, especially in life-altering situations such as radiology or robotic surgery.

The available evidence suggests that frequent engagement with automation induces skill decay. Given that (a) AI is often designed to take over more advanced cognitive processes than non-AI automation, and (b) skill decay is accelerated for cognitive skills, AI-induced skill decay is a likely consequence of frequent engagement with AI assistants.

Depending on the field, AI-induced skill decay may only be a concern in the short-term. If AI assistants can be developed such that they are near perfect in all (or nearly all) circumstances in the future, then developing and maintaining human skills for that task may be unnecessary. But, in fields where problems rapidly evolve or novel events are likely to occur, such as in medicine, or if the AI tool is biased, offline, or making errors, then developing and maintaining human skills will continue to be advantageous.

We propose several lines of inquiry (Figs. 1 and 2) that researchers from multiple perspectives should seek to address related to potential AI-induced skill decay and AI-induced skill development hindrance. The proposed research perspectives are not intended to be discrete areas or mutually exclusive. Likewise, the research questions we propose are not intended to be exhaustive or necessarily the most urgent or important but offer an initial sample of potential research lines in these areas.

Fig. 1.

Fig. 1

Suggested questions for research on AI-induced skill decay

Fig. 2.

Fig. 2

Suggested questions for research on AI-induced skill development hindrance

Further, when addressing these questions, researchers should consider the degree of awareness experts and learners may (or may not) have on their level of skill. If AI assistants lead learners toward an illusion of understanding of the task (Messeri & Crockett, 2024) or lead experts to believe their skills are at a higher level than they are in reality, then basic research, technical research, and applied research are needed to understand the sources, moderators, and solutions to these issues. Only by drawing on multiple disciplinary perspectives can we best understand the impacts of artificial intelligence on humans’ cognitive skills and the future of workplace expertise.

Open practices statement

This paper discusses a perspective on an emerging potential problem and recommends courses of action. We do not report on any studies or original data. Thus, the availability of data and/or materials is not applicable. Likewise, no experiments were conducted and, therefore, preregistration was not applicable.

Acknowledgements

None.

Significance Statement

Artificial intelligence assistants are becoming increasingly common in the workplace. These tools are trained to make high-level decisions that were previously the domain of experts alone. While there are many benefits that come with the use of artificial intelligence assistants, one potential detriment has been largely overlooked: how reliance on artificial intelligence assistants might negatively affect the user’s cognitive skills. Research on the use of automated systems, such as the autopilot in aviation, suggests that their consistent use leads to skill decay among users. Notably, such automated systems are far less sophisticated than artificial intelligence; thus, one might hypothesize that the skill decay observed with these less sophisticated systems is exacerbated with artificial intelligence. We discuss the potential for artificial intelligence assistants to accelerate skill decay among experts and hinder learning during skill acquisition, without the user’s awareness.

Author contributions

BNM drafted the manuscript. All other authors provided feedback and edits. All authors read and approved the final manuscript.

Funding

This material is based upon work supported by the National Science Foundation under Grant No. SES-2129072.

Availability of data and materials

Not applicable.

Declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Footnotes

1

Skill decay of cognitive-oriented tasks may also exhibit varying decay patterns depending on their complexity (see, e.g., Villado et al., 2013).

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  1. Ackerman PL. Individual differences in information processing: An investigation of intellectual abilities and task performance during practice. Intelligence. 1986;10:101–139. doi: 10.1016/0160-2896(86)90010-3. [DOI] [Google Scholar]
  2. Ackerman PL. Determinants of individual differences during skill acquisition: Cognitive abilities and information processing. Journal of Experimental Psychology: General. 1988;117(3):288–318. doi: 10.1037/0096-3445.117.3.288. [DOI] [Google Scholar]
  3. Ackerman PL, Tatel CE. Resolving problems with the skill retention literature: An empirical demonstration and recommendations for researchers. Journal of Experimental Psychology: Applied. 2023 doi: 10.1037/xap0000503. [DOI] [PubMed] [Google Scholar]
  4. Ackerman PL, Woltz DJ. Determinants of learning and performance in an associative memory/substitution task: Task constraints, individual differences, volition, and motivation. Journal of Educational Psychology. 1994;86:487–515. doi: 10.1037/0022-0663.86.4.487. [DOI] [Google Scholar]
  5. Arthur W, Jr, Bennett W, Jr, Stanush PL, McNelly TL. Factors that influence skill decay and retention: A quantitative review and analysis. Human Performance. 1998;11(1):57–101. doi: 10.1207/s15327043hup1101_3. [DOI] [Google Scholar]
  6. Arthur W, Day EA, Bennett W, Portrey AM, editors. Individual and team skill decay: The science and implications for practice. Routledge; 2013. [Google Scholar]
  7. Bajic D, Rickard TC. Toward a generalized theory of the shift to retrieval in cognitive skill learning. Memory & Cognition. 2011;39:1147–1161. doi: 10.3758/s13421-011-0114-z. [DOI] [PubMed] [Google Scholar]
  8. Bankins S, Ocampo AC, Marrone M, Restubog SLD, Woo SE. A multilevel review of artificial intelligence in organizations: Implications for organizational behavior research and practice. Journal of Organizational Behavior. 2023;45(2):159–182. doi: 10.1002/job.2735. [DOI] [Google Scholar]
  9. Barker R. 2030—The future of medicine: Avoiding a medical meltdown. Oxford University Press; 2011. [Google Scholar]
  10. Baron JM, Huang R, McEvoy D, Dighe AS. Use of machine learning to predict clinical decision support compliance, reduce alert burden, and evaluate duplicate laboratory test ordering alerts. JAMIA Open. 2021;4(1):ooab006. doi: 10.1093/jamiaopen/ooab006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Brynjolfsson, E. (2023). A call to augment—not automate—workers. In Generative AI: Perspectives from Stanford HAI. Stanford University Human-Centered Artificial Intelligence. Retrieved March 24, 2024, from https://hai.stanford.edu/sites/default/files/2023-03/Generative_AI_HAI_Perspectives.pdf
  12. Burgoyne AP, Harris LJ, Hambrick DZ. Predicting piano skill acquisition in beginners: The role of general intelligence, music aptitude, and mindset. Intelligence. 2019;76:101383. doi: 10.1016/j.intell.2019.101383. [DOI] [Google Scholar]
  13. Cohen SA. A review of demographic and infrastructural factors and potential solutions to the physician and nursing shortage predicted to impact the growing US elderly population. Journal of Public Health Management and Practice. 2009;15(4):352–362. doi: 10.1097/phh.0b013e31819d817d. [DOI] [PubMed] [Google Scholar]
  14. Casner SM, Geven RW, Recker MP, Schooler JW. The retention of manual flying skills in the automated cockpit. Human Factors. 2014;56(8):1506–1516. doi: 10.1177/0018720814535628. [DOI] [PubMed] [Google Scholar]
  15. Coratti A, Annecchiarico M. Robotics in general surgery: Current status and critical review. OA Robotic Surgery. 2013 doi: 10.13172/2053-3225-1-1-709. [DOI] [Google Scholar]
  16. Dahmani L, Bohbot VD. Habitual use of GPS negatively impacts spatial memory during self-guided navigation. Scientific Reports. 2020;10(1):6310. doi: 10.1038/s41598-020-62877-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Day EA, Arthur W, Villado AJ, Boatman PR, Kowollik V, Bhupatkar A, Bennett W. Relating individual differences in ability, personality, and motivation to the retention and transfer of skill on a complex command-and-control simulation task. In: Arthur W, Day EA, Bennett W, Portray A, editors. Individual and team skill decay. Routlege; 2013. [Google Scholar]
  18. Ebbatson M, Harris D, Huddlestone J, Sears R. The relationship between manual handling performance and recent flying experience in air transport pilots. Ergonomics. 2010;53(2):268–277. doi: 10.1080/00140130903342349. [DOI] [PubMed] [Google Scholar]
  19. Federal Aviation Administration. (2013). Safety alert for operators: Manual flight operations. SAFO 13002. Retrieved March 24, 2024, from https://www.faa.gov/sites/faa.gov/files/other_visit/aviation_industry/airline_operators/airline_safety/SAFO13002.pdf
  20. Federal Aviation Administration. (2022). Advisory circular: Flightpath management. Advisory circular number 120–123. Retrieved February 29, 2024, from https://www.faa.gov/documentLibrary/media/Advisory_Circular/AC_120-123.pdf
  21. Fiorini P, Goldberg KY, Liu Y, Taylor RH. Concepts and trends in autonomy for robot-assisted surgery. Proceedings of the IEEE. 2022;110:993–1011. doi: 10.1109/JPROC.2022.3176828. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Fitts PM, Posner MI. Human performance. Brooks/Cole; 1967. [Google Scholar]
  23. Frank DJ, Demaree HA, Macnamara BN. The role of individual differences in risk learning: Who learns to place optimal wagers? Learning and Motivation. 2020;71:101633. doi: 10.1016/j.lmot.2020.101633. [DOI] [Google Scholar]
  24. Frank DJ, Macnamara BN. How do task characteristics affect learning and performance? The roles of simultaneous, interactive, and continuous tasks. Psychological Research Psychologische Forschung. 2021;85:2364–2397. doi: 10.1007/s00426-020-01394-1. [DOI] [PubMed] [Google Scholar]
  25. Ghaith GM. The relationship between cooperative learning, perception of social support, and academic achievement. System. 2002;30(3):263–273. doi: 10.1016/S0346-251X(02)00014-3. [DOI] [Google Scholar]
  26. Gomes P. Surgical robotics: Reviewing the past, analysing the present, imagining the future. Robotics and Computer-Integrated Manufacturing. 2011;27(2):261–266. doi: 10.1016/j.rcim.2010.06.009. [DOI] [Google Scholar]
  27. Goodman JS, Wood RE. Feedback specificity, learning opportunities, and learning. Journal of Applied Psychology. 2004;89(5):809–821. doi: 10.1037/0021-9010.89.5.809. [DOI] [PubMed] [Google Scholar]
  28. Grier DA. When computers were human. Princeton University Press; 2005. [Google Scholar]
  29. Haslbeck A, Hoerman H-J. Flying the needles: Flight deck automation erodes fine-motor flying skills among airline pilots. Human Factors. 2016;58(4):533–545. doi: 10.1177/0018720816640394. [DOI] [PubMed] [Google Scholar]
  30. Hoffman RR, Ward P, Feltovich PJ, DiBello L, Fiore SM, Andrews DH. Accelerated expertise: Training for high proficiency in a complex world. Psychology Press; 2014. [Google Scholar]
  31. Hosny A, Parmer C, Quuackenbush J, Schwartz LH, Aerts HJWL. Artificial intelligence in radiology. Nature Reviews Cancer. 2018;18:500–518. doi: 10.1038/s41568-018-0016-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Kim JW, Ritter FE, Koubek RJ. An integrated theory for improved skill acquisition and retention in the three stages of learning. Theoretical Issues in Ergonomics Science. 2013;14(1):22–37. doi: 10.1080/1464536X.2011.573008. [DOI] [Google Scholar]
  33. Kluge A, Frank B. Counteracting skill decay: Four refresher interventions and their effect on skill and knowledge retention in a simulated process control task. Ergonomics. 2014;57(2):175–190. doi: 10.1080/00140139.2013.869357. [DOI] [PubMed] [Google Scholar]
  34. Krupinski, E. A. (2000). Medical image perception: Evaluating the role of experience. In Human Vision and Electronic Imaging V, Proceedings Volume 3959, 281–289). 10.1117/12.387164
  35. Leahy, J., & Fisher, A. (2023). FAA shifts focus to pilot manual handling skills. Royal Aeronautical Society. Retrieved February 29, 2024, from https://www.aerosociety.com/news/faa-shifts-focus-to-pilot-manual-handling-skills/
  36. Lechner CM, Miyamoto A, Knopf T. Should students be smart, curious, or both? Fluid intelligence, openness, and interest co-shape the acquisition of reading and math competence. Intelligence. 2019;76:101378. doi: 10.1016/j.intell.2019.101378. [DOI] [Google Scholar]
  37. Logan GD. Toward an instance theory of automatization. Psychological Review. 1988;95:492–527. doi: 10.1037/0033-295X.95.4.492. [DOI] [Google Scholar]
  38. Macnamara BN, Frank DJ. How do task characteristics affect learning and performance? The roles of variably mapped and dynamic tasks. Journal of Experimental Psychology: Learning, Memory, and Cognition. 2018;44(5):764–778. doi: 10.1037/xlm0000487. [DOI] [PubMed] [Google Scholar]
  39. Macnamara BN, Hambrick DZ, Oswald FL. Deliberate practice and performance in music, games, sports, education, and professions: A meta-analysis. Psychological Science. 2014;25:1608–1618. doi: 10.1177/0956797614535810. [DOI] [PubMed] [Google Scholar]
  40. Macnamara BN, Moore AB, Kegl JA, Conway AR. Domain-general cognitive abilities and simultaneous interpreting skill. Interpreting. 2011;13(1):121–142. doi: 10.1075/intp.13.1.08mac. [DOI] [Google Scholar]
  41. Macnamara BN, Prather RW, Burgoyne AP. Beliefs about success are prone to cognitive fallacies. Nature Reviews Psychology. 2023;2:716–717. doi: 10.1038/s44159-023-00255-z. [DOI] [Google Scholar]
  42. Magrabi F, Ammenwerth E, McNair JB, De Keizer NF, Hyppönen H, Nykänen P, Rigby M, Scott PJ, Vehko T, Wong ZS-Y, Georgiou A. Artificial intelligence in clinical decision support: Challenges for evaluating AI and practical implications. Yearbook of Medical Informatics. 2019;28(01):128–134. doi: 10.1055/s-0039-1677903. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Manzey D, Reichenbach J, Onnasch L. Human performance consequences of automated decision aids: The impact of degree of automation and system experience. Journal of Cognitive Engineering and Decision Making. 2012;6(1):57–87. doi: 10.1177/1555343411433844. [DOI] [Google Scholar]
  44. Meinz EJ, Hambrick DZ. Deliberate practice is necessary but not sufficient to explain individual differences in piano sight-reading skill: The role of working memory capacity. Psychological Science. 2010;21(7):914–919. doi: 10.1177/0956797610373933. [DOI] [PubMed] [Google Scholar]
  45. Meinz EJ, Hambrick DZ, Hawkins CB, Gillings AK, Meyer BE, Schneider JL. Roles of domain knowledge and working memory capacity in components of skill in Texas Hold’Em poker. Journal of Applied Research in Memory and Cognition. 2012;1(1):34–40. doi: 10.1016/j.jarmac.2011.11.001. [DOI] [Google Scholar]
  46. Messeri L, Crockett MJ. Artificial intelligence and illusions of understanding in scientific research. Nature. 2024;627(8002):49–58. doi: 10.1038/s41586-024-07146-0. [DOI] [PubMed] [Google Scholar]
  47. Mosier KL, Skitka LJ, Heers S, Burdick M. Automation bias: Decision making and performance in high-tech cockpits. International Journal of Aviation Psychology. 1997 doi: 10.1207/s15327108ijap0801_3. [DOI] [PubMed] [Google Scholar]
  48. Nestojko JF, Finley JR, Roediger HL., III Extending cognition to external agents. Psychological Inquiry. 2014;24(4):321–325. doi: 10.1080/1047840X.2013.844056. [DOI] [Google Scholar]
  49. Panesar S, Cagle Y, Chander D, Morey J, Fernandez-Miranda J, Kliot M. Artificial intelligence and the future of surgical robotics. Annals of Surgery. 2019;270(2):223–226. doi: 10.1097/SLA.0000000000003262. [DOI] [PubMed] [Google Scholar]
  50. Parasuraman R, Manzey DH. Complacency and bias in human use of automation: An attentional integration. Human Factors. 2010;52(3):381–410. doi: 10.1177/0018720810376. [DOI] [PubMed] [Google Scholar]
  51. Rezazade Mehrizi MH, Mol F, Peter M, Ranschaert E, Dos Santos DP, Shahidi R, Fatehi M, Dratsch T. The impact of AI suggestions on radiologists’ decisions: A pilot study of explainability and attitudinal priming interventions in mammography examination. Nature Scientific Reports. 2023;13(1):9230. doi: 10.1038/s41598-023-36435-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Risko EF, Gilbert SJ. Cognitive offloading. Trends in Cognitive Sciences. 2016;20(9):676–688. doi: 10.1016/j.tics.2016.07.002. [DOI] [PubMed] [Google Scholar]
  53. Roediger HL, III, Karpicke JD. Test-enhanced learning: Taking memory tests improves long-term retention. Psychological Science. 2006;17(3):249–255. doi: 10.1111/j.1467-9280.2006.01693.x. [DOI] [PubMed] [Google Scholar]
  54. Schneider W, Shiffrin RM. Controlled and automatic human information processing: I. Detection, search, and attention. Psychological Review. 1977;84(1):1–66. doi: 10.1037/0033-295X.84.1.1. [DOI] [Google Scholar]
  55. Shiffrin RM, Schneider W. Controlled and automatic human information processing: II. Perceptual learning, automatic attending and a general theory. Psychological Review. 1977;84(2):127–190. doi: 10.1037/0033-295X.84.2.127. [DOI] [Google Scholar]
  56. Smith, P. J., & Baumann, E. (2019). Human-automation teaming: Unintended impacts and mitigations for degraded NextGen operations. https://rosap.ntl.bts.gov/view/dot/43777
  57. Smith PJ, McCoy E, Layton C. Brittleness in the design of cooperative problem-solving systems: The effects on user performance. IEEE Transactions on Systems, Man and Cybernetics. 1997;27:360–371. doi: 10.1109/3468.568744. [DOI] [Google Scholar]
  58. Smith SM, Glenberg A, Bjork RA. Environmental context and human memory. Memory & Cognition. 1978;6(4):342–353. doi: 10.3758/BF03197465. [DOI] [Google Scholar]
  59. Villado AJ, Day EA, Arthur W, Boatman PR, Kowollik V, Bhupatkar A, Bennett W. Complex command-and-control simulation task performance following periods of nonuse. In: Arthur W, Day EA, Bennett W, Portray A, editors. Individual and team skill decay. Routlege; 2013. [Google Scholar]
  60. Volz K, Yang E, Dudley R, Lynch E, Dropps M, Dorneich M. An evaluation of cognitive skill degradation in information automation. Proceedings of the Human Factors and Ergonomics Society Annual Meeting. 2016;60(1):191–195. doi: 10.1177/1541931213601043. [DOI] [Google Scholar]
  61. Walsh, T., Levy, N., Bell, G., Elliott, A., Maclaurin, J., Mareels, I., & Wood, F. M. (2019). The effective and ethical development of artificial intelligence: An opportunity to improve our wellbeing. Australian Council of Learned Academies. Retrieved March 24, 2024, from https://acola.org/wp-content/uploads/2019/07/hs4_artificial-intelligence-report.pdf
  62. Wang, X., Day, E. A., Kowollik, V., Schuelke, M. J., & Hughes, M. G. (2013). Factors influencing knowledge and skill decay after training: A meta-analysis. In Individual and Team Skill Decay (pp. 68–116). Routledge.
  63. Wickens CD, Dixon SR. The benefits of imperfect diagnostic automation: A synthesis of the literature. Theoretical Issues in Ergonomics Science. 2007;8(3):201–212. doi: 10.1080/14639220500370105. [DOI] [Google Scholar]
  64. Yang GZ, Cambias J, Cleary K, Daimler E, Drake J, Dupont PE, Hata N, Kazanzides P, Martel S, Patel RV, Santos VJ, Taylor RH. Medical robotics—Regulatory, ethical, and legal considerations for increasing levels of autonomy. Science Robotics. 2017;2(4):eaam8638. doi: 10.1126/scirobotics.aam8638. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Not applicable.


Articles from Cognitive Research: Principles and Implications are provided here courtesy of Springer

RESOURCES